(c) 2007 IUPUI SPEA K300 (4392) Outline Least Squares Methods Estimation: Least Squares Interpretation of estimators Properties of OLS estimators Variance.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Multiple Regression Analysis
Multiple Regression W&W, Chapter 13, 15(3-4). Introduction Multiple regression is an extension of bivariate regression to take into account more than.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Linear regression models
EPI 809/Spring Probability Distribution of Random Error.
Simple Linear Regression and Correlation
Objectives (BPS chapter 24)
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 13 Multiple Regression
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
Multiple regression analysis
Chapter 10 Simple Regression.
9. SIMPLE LINEAR REGESSION AND CORRELATION
Chapter 12 Simple Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 12 Multiple Regression
Chapter 13 Introduction to Linear Regression and Correlation Analysis
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
The Simple Regression Model
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Chapter 11 Multiple Regression.
Linear Regression Example Data
SIMPLE LINEAR REGRESSION
Ch. 14: The Multiple Regression Model building
BCOR 1020 Business Statistics
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Introduction to Regression Analysis, Chapter 13,
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 13-1 Chapter 13 Introduction to Multiple Regression Statistics for Managers.
(c) 2007 IUPUI SPEA K300 (4392) Outline Correlation and Covariance Bivariate Correlation Coefficient Types of Correlation Correlation Coefficient Formula.
Regression and Correlation Methods Judy Zhong Ph.D.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Chapter 13: Inference in Regression
Regression Analysis (2)
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Simple Linear Regression Models
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Lecture 10: Correlation and Regression Model.
Environmental Modeling Basic Testing Methods - Statistics III.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
(c) 2007 IUPUI SPEA K300 (4392) Outline Comparing Group Means Data arrangement Linear Models and Factor Analysis of Variance (ANOVA) Partitioning Variance.
11-1 Copyright © 2014, 2011, and 2008 Pearson Education, Inc.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
The simple linear regression model and parameter estimation
Chapter 20 Linear and Multiple Regression
Correlation and Simple Linear Regression
Simple Linear Regression
Correlation and Simple Linear Regression
Simple Linear Regression
Simple Linear Regression and Correlation
Simple Linear Regression
Presentation transcript:

(c) 2007 IUPUI SPEA K300 (4392) Outline Least Squares Methods Estimation: Least Squares Interpretation of estimators Properties of OLS estimators Variance of Y, b, and a Hypothesis Test of b and a ANOVA table Goodness-of-Fit and R 2

(c) 2007 IUPUI SPEA K300 (4392) Linear regression model

(c) 2007 IUPUI SPEA K300 (4392) Terminology Dependent variable (DV) = response variable = left-hand side (LHS) variable Independent variables (IV) = explanatory variables = right-hand side (RHS) variables = regressor (excluding a or b 0 ) a (b 0 ) is an estimator of parameter α, β 0 b (b 1 ) is an estimator of parameter β, β 1 a and b are the intercept and slope

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method How to draw such a line based on data points observed? Suppose a imaginary line of y= a + bx Imagine a vertical distance (or error) between the line and a data point. E=Y-E(Y) This error (or gap) is the deviation of the data point from the imaginary line, regression line What is the best values of a and b? A and b that minimizes the sum of such errors (deviations of individual data points from the line)

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method Deviation does not have good properties for computation Why do we use squares of deviation? (e.g., variance) Let us get a and b that can minimize the sum of squared deviations rather than the sum of deviations. This method is called least squares

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method Least squares method minimizes the sum of squares of errors (deviations of individual data points form the regression line) Such a and b are called least squares estimators (estimators of parameters α and β). The process of getting parameter estimators (e.g., a and b) is called estimation “Regress Y on X” Lest squares method is the estimation method of ordinary least squares (OLS)

(c) 2007 IUPUI SPEA K300 (4392) Ordinary Least Squares Ordinary least squares (OLS) = Linear regression model = Classical linear regression model Linear relationship between Y and Xs Constant slopes (coefficients of Xs) Least squares method Xs are fixed; Y is conditional on Xs Error is not related to Xs Constant variance of errors

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method 1 How to get a and b that can minimize the sum of squares of errors?

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method 2 Linear algebraic solution Compute a and b so that partial derivatives with respect to a and b are equal to zero

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method 3 Take a partial derivative with respect to b and plug in a you got, a=Ybar –b*Xbar

(c) 2007 IUPUI SPEA K300 (4392) Least Squares Method 4 Least squares method is an algebraic solution that minimizes the sum of squares of errors (variance component of error) Not recommended

(c) 2007 IUPUI SPEA K300 (4392) OLS: Example 10-5 (1) Noxyx-xbary-ybar(x-xb)(y-yb)(x-xbar)^ Mean Sum

(c) 2007 IUPUI SPEA K300 (4392) OLS: Example 10-5 (2), NO! Noxyxyx^ Mean Sum

(c) 2007 IUPUI SPEA K300 (4392) OLS: Example 10-5 (3)

(c) 2007 IUPUI SPEA K300 (4392) What Are a and b ? a is an estimator of its parameter α a is the intercept, a point of y where the regression line meets the y axis b is an estimator of its parameter β b is the slope of the regression line b is constant regardless of values of Xs b is more important than a since that is what researchers want to know.

(c) 2007 IUPUI SPEA K300 (4392) How to interpret b? For unit increase in x, the expected change in y is b, holding other things (variables) constant. For unit increase in x, we expect that y increases by b, holding other things (variables) constant. For unit increase in x, we expect that y increases by.964, holding other variables constant.

(c) 2007 IUPUI SPEA K300 (4392) Properties of OLS estimators The outcome of least squares method is OLS parameter estimators a and b. OLS estimators are linear OLS estimators are unbiased (precise) OLS estimators are efficient (small variance) Gauss-Markov Theorem: Among linear unbiased estimators, least square estimator (OLS estimator) has minimum variance.  BLUE (best linear unbiased estimator)

(c) 2007 IUPUI SPEA K300 (4392) Hypothesis Test of a an b How reliable are a and b we compute? T-test (Wald test in general) can answer The standardized effect size (effect size / standard error) Effect size is a-0 and b-0 assuming 0 is the hypothesized value; H 0 : α=0, H 0 : β=0 Degrees of freedom is N-K, where K is the number of regressors +1 How to compute standard error (deviation)?

(c) 2007 IUPUI SPEA K300 (4392) Variance of b (1) b is a random variable that changes across samples. b is a weighted sum of linear combinations of random variable Y

(c) 2007 IUPUI SPEA K300 (4392) Variance of b (2) Variance of Y (error) is σ 2 Var(kY) = k 2 Var(Y) = k 2 σ 2

(c) 2007 IUPUI SPEA K300 (4392) Variance of a a=Ybar + b*Xbar Var(b)=σ 2 /SS x, SS x = ∑(X-Xbar) 2 Var(∑Y)=Var(Y 1 )+Var(Y 2 )+…+Var(Y n )=nσ 2 Now, how do we compute the variance of Y, σ 2 ?

(c) 2007 IUPUI SPEA K300 (4392) Variance of Y or error Variance of Y is based on residuals (errors), Y-Yhat “Hat” means an estimator of the parameter Y hat is predicted (by a + bX) value of Y; plug in x given a and b to get Y hat Since a regression model includes K parameters (a and b in simple regression), the degrees of freedom is N-K Numerator is SSE in the ANOVA table

(c) 2007 IUPUI SPEA K300 (4392) Illustration (1) Noxyx-xbary-ybar(x-xb)(y-yb)(x-xbar)^2yhat(y-yhat)^ Mean Sum SSE= , MSE=

(c) 2007 IUPUI SPEA K300 (4392) Illustration (2): Test b How to test whether beta is zero (no effect)? Like y, α and β follow a normal distribution; a and b follows the t distribution b=.9644, SE(b)=.2381,df=N-K=6-2=4 Hypothesis Testing 1. H 0 :β=0 (no effect), H a :β≠0 (two-tailed) 2. Significance level=.05, CV=2.776, df=6-2=4 3. TS=( )/.2381=4.0510~t(N-K) 4. TS (4.051)>CV (2.776), Reject H 0 5. Beta (not b) is not zero. There is a significant impact of X on Y

(c) 2007 IUPUI SPEA K300 (4392) Illustration (3): Test a How to test whether alpha is zero? Like y, α and β follow a normal distribution; a and b follows the t distribution a= , SE(a)= , df=N-K=6-2=4 Hypothesis Testing 1. H 0 :α=0, H a :α≠0 (two-tailed) 2. Significance level=.05, CV= TS=( )/ =5.8388~t(N-K) 4. TS (5.839)>CV (2.776), Reject H 0 5. Alpha (not a) is not zero. The intercept is discernable from zero (significant intercept).

(c) 2007 IUPUI SPEA K300 (4392) Questions How do we test H 0 : β 0 (α)=β 1 =β 2 …=0? Remember that t-test compares only two group means, while ANOVA compares more than two group means simultaneously. The same thing in linear regression. Construct the ANOVA table by partitioning variance of Y; F test examines the above H 0 The ANOVA table provides key information of a regression model

(c) 2007 IUPUI SPEA K300 (4392) Partitioning Variance of Y (1)

(c) 2007 IUPUI SPEA K300 (4392) Partitioning Variance of Y (2)

(c) 2007 IUPUI SPEA K300 (4392) Partitioning Variance of Y (3) X Noxyyhat(y-ybar)^2(yhat-ybar)^2(y-yhat)^ Mean SSTSSMSSE Sum =81+.96×43, 148.6= ×70 SST=SSM+SSE, 649.5=

(c) 2007 IUPUI SPEA K300 (4392) ANOVA Table H 0 : all parameters are zero, β 0 = β 1 = 0 H a : at least one parameter is not zero CV is (1,4), TS>CV, reject H 0 SourcesSum of SquaresDFMean SquaresF ModelSSMK-1MSM=SSM/(K-1)MSM/MSE ResidualSSEN-KMSE=SSE/(N-K) TotalSSTN-1 SourcesSum of SquaresDFMean SquaresF Model Residual Total

(c) 2007 IUPUI SPEA K300 (4392) R 2 and Goodness-of-fit Goodness-of-fit measures evaluates how well a regression model fits the data The smaller SSE, the better fit the model F test examines if all parameters are zero. (large F and small p-value indicate good fit) R 2 (Coefficient of Determination) is SSM/SST that measures how much a model explains the overall variance of Y. R 2 =SSM/SST=522.2/649.5=.80 Large R square means the model fits the data

(c) 2007 IUPUI SPEA K300 (4392) Myth and Misunderstanding in R 2 R square is Karl Pearson correlation coefficient squared. r 2 = =.80 If a regression model includes many regressors, R 2 is less useful, if not useless. Addition of any regressor always increases R 2 regardless of the relevance of the regressor Adjusted R 2 give penalty for adding regressors, Adj. R 2 =1-[(N-1)/(N-K)](1-R 2 ) R 2 is not a panacea although its interpretation is intuitive; if the intercept is omitted, R 2 is incorrect. Check specification, F, SSE, and individual parameter estimators to evaluate your model; A model with smaller R 2 can be better in some cases.

(c) 2007 IUPUI SPEA K300 (4392) Interpolation and Extrapolation Confidence interval of E(Y|X), where x is within the rage of data x; interpolation Confidence interval of Y|X, where x is beyond the range of data x; extrapolation Extrapolation involves penalty and danger, which widens the confidence interval; less reliable