12 Multiple Linear Regression CHAPTER OUTLINE

Slides:



Advertisements
Similar presentations
3.3 Hypothesis Testing in Multiple Linear Regression
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
CHAPTER 2 Building Empirical Model. Basic Statistical Concepts Consider this situation: The tension bond strength of portland cement mortar is an important.
Probability & Statistical Inference Lecture 9
12-1 Multiple Linear Regression Models Introduction Many applications of regression analysis involve situations in which there are more than.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Ch11 Curve Fitting Dr. Deshi Ye
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Multiple Linear Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
1 Chapter 9 Variable Selection and Model building Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
8 Statistical Intervals for a Single Sample CHAPTER OUTLINE
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
13-1 Designing Engineering Experiments Every experiment involves a sequence of activities: Conjecture – the original hypothesis that motivates the.
Multiple Linear Regression
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
13 Design and Analysis of Single-Factor Experiments:
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Correlation & Regression
1 14 Design of Experiments with Several Factors 14-1 Introduction 14-2 Factorial Experiments 14-3 Two-Factor Factorial Experiments Statistical analysis.
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 11 Regression.
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Chapter 12 Multiple Regression and Model Building.
Learning Objective Chapter 14 Correlation and Regression Analysis CHAPTER fourteen Correlation and Regression Analysis Copyright © 2000 by John Wiley &
Chapter 12 Multiple Linear Regression Doing it with more variables! More is better. Chapter 12A.
CHAPTER 14 MULTIPLE REGRESSION
6-3 Multiple Regression Estimation of Parameters in Multiple Regression.
● Final exam Wednesday, 6/10, 11:30-2:30. ● Bring your own blue books ● Closed book. Calculators and 2-page cheat sheet allowed. No cell phone/computer.
1 Chapter 3 Multiple Linear Regression Multiple Regression Models Suppose that the yield in pounds of conversion in a chemical process depends.
1 10 Statistical Inference for Two Samples 10-1 Inference on the Difference in Means of Two Normal Distributions, Variances Known Hypothesis tests.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Copyright © 2010 Pearson Addison-Wesley. All rights reserved. Chapter 12 Multiple Linear Regression and Certain Nonlinear Regression Models.
Linear Regression Analysis 5E Montgomery, Peck & Vining 1 Chapter 8 Indicator Variables.
6-3 Multiple Regression Estimation of Parameters in Multiple Regression.
1 9 Tests of Hypotheses for a Single Sample. © John Wiley & Sons, Inc. Applied Statistics and Probability for Engineers, by Montgomery and Runger. 9-1.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Simple Linear Regression (SLR)
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Chap 13-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 13 Multiple Regression and.
Regression Analysis1. 2 INTRODUCTION TO EMPIRICAL MODELS LEAST SQUARES ESTIMATION OF THE PARAMETERS PROPERTIES OF THE LEAST SQUARES ESTIMATORS AND ESTIMATION.
Simple linear regression and correlation Regression analysis is the process of constructing a mathematical model or function that can be used to predict.
Statistics 350 Review. Today Today: Review Simple Linear Regression Simple linear regression model: Y i =  for i=1,2,…,n Distribution of errors.
Chapter Outline EMPIRICAL MODELS 11-2 SIMPLE LINEAR REGRESSION 11-3 PROPERTIES OF THE LEAST SQUARES ESTIMATORS 11-4 SOME COMMENTS ON USES OF REGRESSION.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Applied Statistics and Probability for Engineers
Chapter 12 Multiple Linear Regression and Certain Nonlinear Regression Models.
Chapter 15 Multiple Regression and Model Building
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Multiple Linear Regression
Slides by JOHN LOUCKS St. Edward’s University.
6-1 Introduction To Empirical Models
Prepared by Lee Revere and John Large
Chapter 3 Multiple Linear Regression
Simple Linear Regression
Multivariate Linear Regression
Essentials of Statistics for Business and Economics (8e)
Model Adequacy Checking
Presentation transcript:

12 Multiple Linear Regression CHAPTER OUTLINE 12-1 Multiple Linear Regression Model 12-3 Confidence Intervals in Multiple Linear Regression 12-1.1 Introduction 12-1.2 Least squares estimation of the parameters 12-4.1 Use of t-tests 12-3.2 Confidence interval on the mean response 12-1.3 Matrix approach to multiple linear regression 12-4 Prediction of New Observations 12-5 Model Adequacy Checking 12-1.4 Properties of the least squares estimators 12-5.1 Residual analysis 12-2 Hypothesis Tests in Multiple Linear Regression 12-5.2 Influential observations 12-6 Aspects of Multiple Regression Modeling 12-2.1 Test for significance of regression 12-2.2 Tests on individual regression coefficients & subsets of coefficients 12-6.1 Polynomial regression models 12-6.2 Categorical regressors & indicator variables 12-6.3 Selection of variables & model building 12-6.4 Multicollinearity Dear Instructor: This file is an adaptation of the 4th edition slides for the 5th edition. It will be replaced as slides are developed following the style of the Chapters 1-7 slides.

Learning Objectives for Chapter 12 After careful study of this chapter, you should be able to do the following: Use multiple regression techniques to build empirical models to engineering and scientific data. Understand how the method of least squares extends to fitting multiple regression models. Assess regression model adequacy. Test hypotheses and construct confidence intervals on the regression coefficients. Use the regression model to estimate the mean response, and to make predictions and to construct confidence intervals and prediction intervals. Build regression models with polynomial terms. Use indicator variables to model categorical regressors. Use stepwise regression and other model building techniques to select the appropriate set of variables for a regression model.

12-1: Multiple Linear Regression Models 12-1.1 Introduction Many applications of regression analysis involve situations in which there are more than one regressor variable. A regression model that contains more than one regressor variable is called a multiple regression model.

12-1: Multiple Linear Regression Models 12-1.1 Introduction For example, suppose that the effective life of a cutting tool depends on the cutting speed and the tool angle. A possible multiple regression model could be where Y – tool life x1 – cutting speed x2 – tool angle

12-1: Multiple Linear Regression Models 12-1.1 Introduction Figure 12-1 (a) The regression plane for the model E(Y) = 50 + 10x1 + 7x2. (b) The contour plot

12-1: Multiple Linear Regression Models 12-1.1 Introduction

12-1: Multiple Linear Regression Models 12-1.1 Introduction Figure 12-2 (a) Three-dimensional plot of the regression model E(Y) = 50 + 10x1 + 7x2 + 5x1x2. (b) The contour plot

12-1: Multiple Linear Regression Models 12-1.1 Introduction Figure 12-3 (a) Three-dimensional plot of the regression model E(Y) = 800 + 10x1 + 7x2 – 8.5x12 – 5x22 + 4x1x2. (b) The contour plot

12-1: Multiple Linear Regression Models 12-1.2 Least Squares Estimation of the Parameters

12-1: Multiple Linear Regression Models 12-1.2 Least Squares Estimation of the Parameters The least squares function is given by The least squares estimates must satisfy

12-1: Multiple Linear Regression Models 12-1.2 Least Squares Estimation of the Parameters The least squares normal Equations are The solution to the normal Equations are the least squares estimators of the regression coefficients.

12-1: Multiple Linear Regression Models Example 12-1

12-1: Multiple Linear Regression Models Example 12-1

12-1: Multiple Linear Regression Models Figure 12-4 Matrix of scatter plots (from Minitab) for the wire bond pull strength data in Table 12-2.

12-1: Multiple Linear Regression Models Example 12-1

12-1: Multiple Linear Regression Models Example 12-1

12-1: Multiple Linear Regression Models Example 12-1

12-1: Multiple Linear Regression Models 12-1.3 Matrix Approach to Multiple Linear Regression Suppose the model relating the regressors to the response is In matrix notation this model can be written as

12-1: Multiple Linear Regression Models 12-1.3 Matrix Approach to Multiple Linear Regression where

12-1: Multiple Linear Regression Models 12-1.3 Matrix Approach to Multiple Linear Regression We wish to find the vector of least squares estimators that minimizes: The resulting least squares estimate is

12-1: Multiple Linear Regression Models 12-1.3 Matrix Approach to Multiple Linear Regression

12-1: Multiple Linear Regression Models Example 12-2

Example 12-2

12-1: Multiple Linear Regression Models Example 12-2

12-1: Multiple Linear Regression Models Example 12-2

12-1: Multiple Linear Regression Models Example 12-2

12-1: Multiple Linear Regression Models Example 12-2

12-1: Multiple Linear Regression Models Estimating 2 An unbiased estimator of 2 is

12-1: Multiple Linear Regression Models 12-1.4 Properties of the Least Squares Estimators Unbiased estimators: Covariance Matrix:

12-1: Multiple Linear Regression Models 12-1.4 Properties of the Least Squares Estimators Individual variances and covariances: In general,

12-2: Hypothesis Tests in Multiple Linear Regression 12-2.1 Test for Significance of Regression The appropriate hypotheses are The test statistic is

12-2: Hypothesis Tests in Multiple Linear Regression 12-2.1 Test for Significance of Regression

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-3

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-3

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-3

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-3

12-2: Hypothesis Tests in Multiple Linear Regression R2 and Adjusted R2 The coefficient of multiple determination For the wire bond pull strength data, we find that R2 = SSR/SST = 5990.7712/6105.9447 = 0.9811. Thus, the model accounts for about 98% of the variability in the pull strength response.

12-2: Hypothesis Tests in Multiple Linear Regression R2 and Adjusted R2 The adjusted R2 is The adjusted R2 statistic penalizes the analyst for adding terms to the model. It can help guard against overfitting (including regressors that are not really useful)

12-2: Hypothesis Tests in Multiple Linear Regression 12-2.2 Tests on Individual Regression Coefficients and Subsets of Coefficients The hypotheses for testing the significance of any individual regression coefficient:

12-2: Hypothesis Tests in Multiple Linear Regression 12-2.2 Tests on Individual Regression Coefficients and Subsets of Coefficients The test statistic is Reject H0 if |t0| > t/2,n-p. This is called a partial or marginal test

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-4

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-4

12-2: Hypothesis Tests in Multiple Linear Regression The general regression significance test or the extra sum of squares method: We wish to test the hypotheses:

12-2: Hypothesis Tests in Multiple Linear Regression A general form of the model can be written: where X1 represents the columns of X associated with 1 and X2 represents the columns of X associated with 2

12-2: Hypothesis Tests in Multiple Linear Regression For the full model: If H0 is true, the reduced model is

12-2: Hypothesis Tests in Multiple Linear Regression The test statistic is: Reject H0 if f0 > f,r,n-p The test in Equation (12-32) is often referred to as a partial F-test

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-6

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-6

12-2: Hypothesis Tests in Multiple Linear Regression Example 12-6

12-3: Confidence Intervals in Multiple Linear Regression 12-3.1 Confidence Intervals on Individual Regression Coefficients Definition

12-3: Confidence Intervals in Multiple Linear Regression Example 12-7

12-3: Confidence Intervals in Multiple Linear Regression 12-3.2 Confidence Interval on the Mean Response The mean response at a point x0 is estimated by The variance of the estimated mean response is

12-3: Confidence Intervals in Multiple Linear Regression 12-3.2 Confidence Interval on the Mean Response Definition

12-3: Confidence Intervals in Multiple Linear Regression Example 12-8

12-3: Confidence Intervals in Multiple Linear Regression Example 12-8

12-4: Prediction of New Observations A point estimate of the future observation Y0 is A 100(1-)% prediction interval for this future observation is

12-4: Prediction of New Observations Figure 12-5 An example of extrapolation in multiple regression

12-4: Prediction of New Observations Example 12-9

12-5: Model Adequacy Checking 12-5.1 Residual Analysis Example 12-10 Figure 12-6 Normal probability plot of residuals

12-5: Model Adequacy Checking 12-5.1 Residual Analysis Example 12-10

12-5: Model Adequacy Checking 12-5.1 Residual Analysis Example 12-10 Figure 12-7 Plot of residuals

12-5: Model Adequacy Checking 12-5.1 Residual Analysis Example 12-10 Figure 12-8 Plot of residuals against x1.

12-5: Model Adequacy Checking 12-5.1 Residual Analysis Example 12-10 Figure 12-9 Plot of residuals against x2.

12-5: Model Adequacy Checking 12-5.1 Residual Analysis

12-5: Model Adequacy Checking 12-5.1 Residual Analysis The variance of the ith residual is

12-5: Model Adequacy Checking 12-5.1 Residual Analysis

12-5: Model Adequacy Checking 12-5.2 Influential Observations Figure 12-10 A point that is remote in x-space.

12-5: Model Adequacy Checking 12-5.2 Influential Observations Cook’s distance measure

12-5: Model Adequacy Checking Example 12-11

12-5: Model Adequacy Checking Example 12-11

12-6: Aspects of Multiple Regression Modeling 12-6.1 Polynomial Regression Models

12-6: Aspects of Multiple Regression Modeling Example 12-12

12-6: Aspects of Multiple Regression Modeling Example 12-11 Figure 12-11 Data for Example 12-11.

Example 12-12

12-6: Aspects of Multiple Regression Modeling Example 12-12

12-6: Aspects of Multiple Regression Modeling 12-6.2 Categorical Regressors and Indicator Variables Many problems may involve qualitative or categorical variables. The usual method for the different levels of a qualitative variable is to use indicator variables. For example, to introduce the effect of two different operators into a regression model, we could define an indicator variable as follows:

12-6: Aspects of Multiple Regression Modeling Example 12-13

12-6: Aspects of Multiple Regression Modeling Example 12-13

12-6: Aspects of Multiple Regression Modeling Example 12-13

Example 12-12

12-6: Aspects of Multiple Regression Modeling Example 12-13

12-6: Aspects of Multiple Regression Modeling Example 12-13

12-6: Aspects of Multiple Regression Modeling 12-6.3 Selection of Variables and Model Building

12-6: Aspects of Multiple Regression Modeling 12-6.3 Selection of Variables and Model Building All Possible Regressions – Example 12-14

12-6: Aspects of Multiple Regression Modeling 12-6.3 Selection of Variables and Model Building All Possible Regressions – Example 12-14

12-6: Aspects of Multiple Regression Modeling 12-6.3 Selection of Variables and Model Building All Possible Regressions – Example 12-14 Figure 12-12 A matrix of Scatter plots from Minitab for the Wine Quality Data.

12-6.3: Selection of Variables and Model Building - Stepwise Regression Example 12-14

12-6.3: Selection of Variables and Model Building - Backward Regression Example 12-14

12-6: Aspects of Multiple Regression Modeling 12-6.4 Multicollinearity Variance Inflation Factor (VIF)

12-6: Aspects of Multiple Regression Modeling 12-6.4 Multicollinearity The presence of multicollinearity can be detected in several ways. Two of the more easily understood of these are:

Important Terms & Concepts of Chapter 12 All possible regressions Analysis of variance test in multiple regression Categorical variables Confidence intervals on the mean response Cp statistic Extra sum of squares method Hidden extrapolation Indicator variables Inference (test & intervals) on individual model parameters Influential observations Model parameters & their interpretation in multiple regression Multicollinearity Multiple regression Outliers Polynomial regression model Prediction interval on a future observation PRESS statistic Residual analysis & model adequacy checking Significance of regression Stepwise regression & related methods Variance Inflation Factor (VIF)