Economics 173 Business Statistics Lecture 20 Fall, 2001© Professor J. Petry

Slides:



Advertisements
Similar presentations
Chapter 18 Multiple Regression.
Advertisements

BA 275 Quantitative Business Methods
Student Quiz Grades Test Grades 1.Describe the association between Quiz Grades and Test Grades. 2.Write the.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
1 Multiple Regression Model Error Term Assumptions –Example 1: Locating a motor inn Goodness of Fit (R-square) Validity of estimates (t-stats & F-stats)
Lecture 9- Chapter 19 Multiple regression Introduction In this chapter we extend the simple linear regression model and allow for any number of.
LECTURE 3 Introduction to Linear Regression and Correlation Analysis
Chapter 13 Multiple Regression
Chapter 12 Simple Regression
Simple Linear Regression
1 Multiple Regression Chapter Introduction In this chapter we extend the simple linear regression model, and allow for any number of independent.
Lecture 25 Multiple Regression Diagnostics (Sections )
Chapter 12 Multiple Regression
Regression Diagnostics Using Residual Plots in SAS to Determine the Appropriateness of the Model.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
1 Multiple Regression. 2 Introduction In this chapter we extend the simple linear regression model, and allow for any number of independent variables.
Lecture 24 Multiple Regression (Sections )
Linear Regression and Correlation Analysis
Lecture 20 Simple linear regression (18.6, 18.9)
Regression Diagnostics - I
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
1 Simple Linear Regression and Correlation Chapter 17.
Lecture 23 Multiple Regression (Sections )
Statistics 350 Lecture 10. Today Last Day: Start Chapter 3 Today: Section 3.8 Homework #3: Chapter 2 Problems (page 89-99): 13, 16,55, 56 Due: February.
Business Statistics - QBM117 Interval estimation for the slope and y-intercept Hypothesis tests for regression.
1 4. Multiple Regression I ECON 251 Research Methods.
BCOR 1020 Business Statistics Lecture 24 – April 17, 2008.
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
1 Simple Linear Regression Chapter Introduction In Chapters 17 to 19 we examine the relationship between interval variables via a mathematical.
Business Statistics - QBM117 Statistical inference for regression.
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
1 Simple Linear Regression 1. review of least squares procedure 2. inference for least squares lines.
Inference for regression - Simple linear regression
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Introduction to Linear Regression
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Regression Analysis Week 8 DIAGNOSTIC AND REMEDIAL MEASURES Residuals The main purpose examining residuals Diagnostic for Residuals Test involving residuals.
Chapter 13 Multiple Regression
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
REGRESSION DIAGNOSTICS Fall 2013 Dec 12/13. WHY REGRESSION DIAGNOSTICS? The validity of a regression model is based on a set of assumptions. Violation.
Economics 173 Business Statistics Lecture 19 Fall, 2001© Professor J. Petry
Lecture 10: Correlation and Regression Model.
Economics 173 Business Statistics Lecture 10 Fall, 2001 Professor J. Petry
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Linear Regression Linear Regression. Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Purpose Understand Linear Regression. Use R functions.
Stat 112 Notes 14 Assessing the assumptions of the multiple regression model and remedies when assumptions are not met (Chapter 6).
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 17 Simple Linear Regression and Correlation.
Economics 173 Business Statistics Lecture 18 Fall, 2001 Professor J. Petry
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
Simple Linear Regression and Correlation (Continue..,) Reference: Chapter 17 of Statistics for Management and Economics, 7 th Edition, Gerald Keller. 1.
Statistics for Business and Economics Module 2: Regression and time series analysis Spring 2010 Lecture 6: Multiple Regression Model Building Priyantha.
1 Assessment and Interpretation: MBA Program Admission Policy The dean of a large university wants to raise the admission standards to the popular MBA.
Quantitative Methods Residual Analysis Multiple Linear Regression C.W. Jackson/B. K. Gordor.
1 Simple Linear Regression Chapter Introduction In Chapters 17 to 19 we examine the relationship between interval variables via a mathematical.
Warm-Up The least squares slope b1 is an estimate of the true slope of the line that relates global average temperature to CO2. Since b1 = is very.
Chapter 14 Introduction to Multiple Regression
Inference for Least Squares Lines
Linear Regression.
Chapter 12: Regression Diagnostics
Prepared by Lee Revere and John Large
Inferences for Regression
Presentation transcript:

Economics 173 Business Statistics Lecture 20 Fall, 2001© Professor J. Petry

2

3 Example – Vacation Homes (18.1) 1.What is the standard error of the estimate? Interpret its value. 2.What is the coefficient of determination? What does this statistic tell you? 3.What is the coefficient of determination, adjusted for degrees of freedom? Why does this value differ from the coefficient of determination? What does this tell you about the model? ========================================================= 1.Test the overall validity of the model. What does the p-value of the test statistic tell you? 2.Interpret each of the coefficients. 3.Test to determine whether each of the independent variables is linearly related to the price of the lot.

4 The required conditions for the model assessment to apply must be checked. –Is the error variable normally distributed? –Is the error variance constant? –Are the errors independent? –Can we identify outliers? –Is multicollinearity a problem? 18.4 Regression Diagnostics - II Draw a histogram of the residuals Plot the residuals versus y ^ Plot the residuals versus the time periods

5 Example 18.2 House price and multicollinearity –A real estate agent believes that a house selling price can be predicted using the house size, number of bedrooms, and lot size. –A random sample of 100 houses was drawn and data recorded. –Analyze the relationship among the four variables

6 Solution The proposed model is PRICE =  0 +  1 BEDROOMS +  2 H-SIZE +  3 LOTSIZE +  –Excel solution The model is valid, but no variable is significantly related to the selling price !!

7 –when regressing the price on each independent variable alone, it is found that each variable is strongly related to the selling price. –Multicollinearity is the source of this problem. Multicollinearity causes two kinds of difficulties: –The t statistics appear to be too small. –The  coefficients cannot be interpreted as “slopes”. However,

8 Remedying violations of the required conditions – Nonnormality or heteroscedasticity can be remedied using transformations on the y variable. –The transformations can improve the linear relationship between the dependent variable and the independent variables. –Many computer software systems allow us to make the transformations easily.

9 A brief list of transformations »y’ = log y (for y > 0) Use when the s  increases with y, or Use when the error distribution is positively skewed »y’ = y 2 Use when the s 2  is proportional to E(y), or Use when the error distribution is negatively skewed »y’ = y 1/2 (for y > 0) Use when the s 2  is proportional to E(y) »y’ = 1/y Use when s 2  increases significantly when y increases beyond some value.

10 Example 18.3: Analysis, diagnostics, transformations. –A statistics professor wanted to know whether time limit affect the marks on a quiz? –A random sample of 100 students was split into 5 groups. –Each student wrote a quiz, but each group was given a different time limit. See data below. MarksMarks Analyze these results, and include diagnostics

11 This model is useful and provides a good fit. The errors seem to be normally distributed The model tested: MARK =  0 +  1 TIME + 

12 The standard error of estimate seems to increase with the predicted value of y. Two transformations are used to remedy this problem: 1. y’ = log e y 2. y’ = 1/y

13 Let us see what happens when a transformation is applied 40,18 40,23 40, , 2.89 Log e 23 = Log e 18 = 2.89 The original data, where “Mark” is a function of “Time” The modified data, where LogMark is a function of “Time"

14 The new regression analysis and the diagnostics are: The model tested: LOGMARK =  ’ 0 +  ’ 1 TIME +  ’ Predicted LogMark = Time This model is useful and provides a good fit.

15 The errors seem to be normally distributed The standard errors still changes with the predicted y, but the change is smaller than before.

16 Let TIME = 55 minutes LogMark = Time = (55) = To find the predicted mark, take the antilog: antilog e = e = How do we use the modified model to predict?