Regression Analysis Part C Confidence Intervals and Hypothesis Testing

Slides:



Advertisements
Similar presentations
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Objectives (BPS chapter 24)
Lecture 8 Relationships between Scale variables: Regression Analysis
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Chapter 10 Simple Regression.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
The Simple Regression Model
SIMPLE LINEAR REGRESSION
Chapter 11 Multiple Regression.
Simple Linear Regression Analysis
1 1 Slide © 2003 South-Western/Thomson Learning™ Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
Chapter 12 Section 1 Inference for Linear Regression.
Simple Linear Regression Analysis
Ordinary Least Squares
Multiple Linear Regression Analysis
Lecture 15 Basics of Regression Analysis
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 11 Regression.
Regression and Correlation Methods Judy Zhong Ph.D.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Inference for regression - Simple linear regression
Hypothesis Testing in Linear Regression Analysis
Regression Method.
CPE 619 Simple Linear Regression Models Aleksandar Milenković The LaCASA Laboratory Electrical and Computer Engineering Department The University of Alabama.
Simple Linear Regression Models
Correlation and Regression
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide © 2004 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Topic 14: Inference in Multiple Regression. Outline Review multiple linear regression Inference of regression coefficients –Application to book example.
Introduction to Linear Regression
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
2014. Engineers often: Regress data  Analysis  Fit to theory  Data reduction Use the regression of others  Antoine Equation  DIPPR We need to be.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Regression Analysis Part B Calculation Procedures Read Chapters 3, 4 and 5 of Forecasting and Time Series, An Applied Approach.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Slide 1 DSCI 5340: Predictive Modeling and Business Forecasting Spring 2013 – Dr. Nick Evangelopoulos Lecture 2: Review of Multiple Regression (Ch. 4-5)
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
Regression Analysis Part A Basic Linear Regression Analysis and Estimation of Parameters Read Chapters 3, 4 and 5 of Forecasting and Time Series, An Applied.
Chapter 13 Simple Linear Regression
Regression Analysis AGEC 784.
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Correlation and Simple Linear Regression
Regression Analysis Part D Model Building
John Loucks St. Edward’s University . SLIDES . BY.
Correlation and Regression
CHAPTER 29: Multiple Regression*
Prepared by Lee Revere and John Large
Correlation and Regression
SIMPLE LINEAR REGRESSION
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Regression Analysis Part C Confidence Intervals and Hypothesis Testing Read Chapters 3, 4 and 5 of Forecasting and Time Series, An Applied Approach. L01C MGS 8110 - Regression Inference

Regression Analysis Modules Part A – Basic Model & Parameter Estimation Part B – Calculation Procedures Part C – Inference: Confidence Intervals & Hypothesis Testing Part D – Goodness of Fit Part E – Model Building Part F – Transformed Variables Part G – Standardized Variables Part H – Dummy Variables Part I – Eliminating Intercept Part J - Outliers Part K – Regression Example #1 Part L – Regression Example #2 Part N – Non-linear Regression Part P – Non-linear Example R L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Overview of Part L01C Confidence Intervals and Hypothesis Testing For Yi prediction and Yi mean Formulas for univariate and multivariate cases. Example calculation: 1) Manual in Excel and 2) SPSS. For Regression Coefficients, bi Example calculation: 1) Data Analysis in Excel and 2) SPSS. Hypothesis Testing For Entire Regression Model, F-test L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Underlying Statistical Theory Underlying Statistical Theory Confidence Intervals and Hypothesis Testing L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

The Standard Error of a Regression Equation The Standard Error of a Regression Equation single independent variable where Yi is the actually observed values of the dependent variable. Yihat is the predicted value from the fitted regression equation. p = 1 is the number of independent variables. k = p+1 = 2 for the number of parameters, b0, b1. n is the sample size used when calculating s. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Individual Prediction Confidence Interval for Individual Prediction single independent variable where f denotes the future (forecasted) or predicted value. p = 1 is the number of independent variables. k = p+1 = 2 for the number of parameters, b0, b1. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Mean Prediction. single independent variable Confidence Interval for Mean Prediction single independent variable (1 of 2) where f denotes the future (forecasted) or predicted value. p = 1 is the number of independent variables. k = p+1 = 2 for the number of parameters, b0, b1. n is the sample size used when calculating s. m is the sample size that is going to be used to calculate the mean value. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Mean Prediction. single independent variable Confidence Interval for Mean Prediction single independent variable (2 of 2) When m=1, the CI for the mean becomes the CI for an individual Y. When m = infinity, the CI for the mean become the CI for a general mean. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

L01C MGS 8110 - Regression Inference

CI Manual Calculations single independent variable L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

CI Manual Calculations single independent variable L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations single independent variable L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations. single independent variable SPSS Data Analysis Calculations single independent variable (continued) L01C MGS 8110 - Regression Inference

The Standard Error of a Regression Equation multivariate case where Y is the actually observed values of the dependent variable, an [n x 1] matrix vector. X is the actually observed values of the independent variable, an [n x 1] matrix vector. b is the calculated regression parameters, a [k x 1] matrix. b=(X’X)-1(X’Y) p is the number of independent variables. k=p+1 is the number of parameters, b0, b1, … bp. n is the sample size used when calculating s. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Individual Predictions multivariate case where Xf is a matrix vector of specified values for the independent variables. X’f = [1 Xf,1, Xf,2, … Xf,p] p is the number of independent variables. k = p+1 is the number of parameters, b0, b1, … bp. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Mean Predictions multivariate case where Xf is a matrix vector of specified values for the independent variables. X’f = [1 Xf,1, Xf,2, … Xf,p] p is the number of independent variables. k = p+1 is the number of parameters, b0, b1, … bp. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

CI Manual Calculations multivariate case L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations multivariate case L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations multivariate case (continued) L01C MGS 8110 - Regression Inference

The Standard Error of a Regression Equation where Yi is the actually observed values of the dependent variable. Yihat is the predicted value from the fitted regression equation. p = 1 is the number of independent variables. k = p+1 = 2 for the number of parameters, b0, b1. n is the sample size used when calculating s. Review from previous slide. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Skip’s Quick and Dirty method to Estimate the Confidence Interval for a Regression Line. Procedure: Select a range of X values from Minimum X to Maximum X. Calculate the corresponding predicted values for Y, Yhat. Add and subtract 2 times the Standard Error for Regression to the predicted values. Optional – plot the two CL line on the scatter plot. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for. Regression Coefficients Confidence Interval for Regression Coefficients single independent variable where p = 1 is the number of independent variables. k = p+1 = 2 for the number of parameters, b0, b1. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Confidence Interval for Regression Coefficients multivariate case where p is the number of independent variables. k = p+1 is the number of parameters, b0, b1, … bp. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Excel, Data Analysis Calculations Multivariate Case L01C MGS 8110 - Regression Inference

Excel, Data Analysis Calculations Multivariate Case (continued) L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations Multivariate Case L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations Multivariate Case (continued) L01C MGS 8110 - Regression Inference

Hypothesis Test of Regression Coefficient where p is the number of independent variables. k = p+1 is the number of parameters, b0, b1, … bp. n is the sample size used when calculating s. 1-a is the confidence level, typically .95. So a/2 = .025. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Excel, Data Analysis Calculation Multivariate Case L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations Multivariate Case L01C MGS 8110 - Regression Inference

Summary: Never test the intercept (constant) Summary: Never test the intercept (constant). Discussed in more detail in L01I If sig is less than .05, keep the variable (slope not equal to zero). If sig is greater than .05, consider eliminating the variable from the model (slope could be zero). L01C MGS 8110 - Regression Inference

Summary: Never test the intercept (constant). If sig is less than Summary: Never test the intercept (constant). If sig is less than .05, keep the variable (slope not equal to zero). If sig is greater than .05, consider eliminating the variable from the model (slope could be zero). If you can’t remember theses rules a year from now, look at the confidence interval. Does the confidence interval contain 0 (zero) L01C MGS 8110 - Regression Inference

F-test for Overall Model L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Excel, Data Analysis Calculation Multivariate Case L01C MGS 8110 - Regression Inference

SPSS Data Analysis Calculations Multivariate Case L01C MGS 8110 - Regression Inference

Review of ANOVA Analysis Green = Residual from mean. Blue, dashed = portion of residual explained by regression equation. Red = portion of residual still unexplained after fitting regression equation. L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Fundamental Concept of ANOVA Analysis Residual Analysis Total = Unexplained + Explained It can be shown (algebraically complex) Total SS = Unexplained SS + Explained SS L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table (1 of 3) Terminology and Table Calculations L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table (2 of 3) Algebraic explanation of terms L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table (3 of 3) Calculation formulas L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table. (1 of 3). Matrix explanation of terms Review of ANOVA Table (1 of 3) Matrix explanation of terms Regression prediction compared to prediction mean of y L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table (2 of 3) Alternative matrix explanation of terms Regression prediction compared to prediction 0 (zero) L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Review of ANOVA Table (3 of 3) Alternative matrix explanation of terms Regression prediction compared to prediction mean of Y & 0 (zero) L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference

Statistical Assumptions 0. The expected value of the residuals is zero, E(ei)=0. The algebraic equation is the correct functional form and accurately predicts E(Yi,j) for all j. Inference Assumptions The residual variance is constant. That is, sj,j2 = s2 for all Xj,j and all i and j. The variance of the observations (Yi,j) does not change as more observations are obtained and/or as different values of Xj are observed. The observations are statistically independent. That is, Yi,j is statistically independent of all other Y’,j values for all i (& j fixed). Knowing the current value of Y does not provide insights into the value of the next Y. The residual errors are normally distributed. The ei,j terms are N(0,s2). L01C MGS 8110 - Regression Inference L01C MGS 8110 - Regression Inference