Multiple Regression Lecture 13 Lecture 12.

Slides:



Advertisements
Similar presentations
Multiple Regression.
Advertisements

Autocorrelation and Heteroskedasticity
Tests of Static Asset Pricing Models
Managerial Economics in a Global Economy
Multiple Regression W&W, Chapter 13, 15(3-4). Introduction Multiple regression is an extension of bivariate regression to take into account more than.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Econ 140 Lecture 81 Classical Regression II Lecture 8.
Quantitative Methods 2 Lecture 3 The Simple Linear Regression Model Edmund Malesky, Ph.D., UCSD.
Chapter 10 Curve Fitting and Regression Analysis
Regression Analysis Using Excel. Econometrics Econometrics is simply the statistical analysis of economic phenomena Here, we just summarize some of the.
Econ 140 Lecture 151 Multiple Regression Applications Lecture 15.
Chapter 15 (Ch. 13 in 2nd Can.) Association Between Variables Measured at the Interval-Ratio Level: Bivariate Correlation and Regression.
Models with Discrete Dependent Variables
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Econometric Details -- the market model Assume that asset returns are jointly multivariate normal and independently and identically distributed through.
Classical Regression III
The Simple Linear Regression Model: Specification and Estimation
Econ 140 Lecture 121 Prediction and Fit Lecture 12.
Chapter 10 Simple Regression.
GG 313 Geological Data Analysis # 18 On Kilo Moana at sea October 25, 2005 Orthogonal Regression: Major axis and RMA Regression.
Econ 140 Lecture 131 Multiple Regression Models Lecture 13.
Chapter 4 Multiple Regression.
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
Multiple Regression Models
Econ 140 Lecture 181 Multiple Regression Applications III Lecture 18.
Econ 140 Lecture 171 Multiple Regression Applications II &III Lecture 17.
Multiple Regression Applications
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Week Lecture 3Slide #1 Minimizing e 2 : Deriving OLS Estimators The problem Deriving b 0 Deriving b 1 Interpreting b 0 and b 1.
Autocorrelation Lecture 18 Lecture 18.
Introduction to Regression Analysis, Chapter 13,
DERIVING LINEAR REGRESSION COEFFICIENTS
Objectives of Multiple Regression
Least-Squares Regression
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Measurement Tools for Science Observation Hypothesis generation Hypothesis testing.
Integrals 5.
MULTIPLE RESTRICTIONS AND ZERO RESTRICTIONS
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Linear Functions 2 Sociology 5811 Lecture 18 Copyright © 2004 by Evan Schofer Do not copy or distribute without permission.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
Specification Error I.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Warm-up Given these solutions below: write the equation of the polynomial: 1. {-1, 2, ½)
Go to Table of Content Single Variable Regression Farrokh Alemi, Ph.D. Kashif Haqqi M.D.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1Spring 02 First Derivatives x y x y x y dy/dx = 0 dy/dx > 0dy/dx < 0.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
January 13, 2014 Multivariate Expressions and Equations.
Chap 6 Further Inference in the Multiple Regression Model
Cross Products and Proportions
Lecture 8: Ordinary Least Squares Estimation BUEC 333 Summer 2009 Simon Woodcock.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
The Simple Linear Regression Model: Specification and Estimation  Theory suggests many relationships between variables  These relationships suggest that.
The simple linear regression model and parameter estimation
§ 4.2 The Exponential Function e x.
Bell Ringer Solve even #’s.
Probability Theory and Parameter Estimation I
Solving Multi-Step Equations
Further Inference in the Multiple Regression Model
Ch6 Dummy Variable Regression Models
CHAPTER 29: Multiple Regression*
Without a calculator, simplify the expressions:
Geology Geomath Chapter 7 - Statistics tom.h.wilson
Introduction Solving inequalities is similar to solving equations. To find the solution to an inequality, use methods similar to those used in solving.
Some issues in multivariate regression
Statistics II: An Overview of Statistics
Concept 5 Rational expressions.
Presentation transcript:

Multiple Regression Lecture 13 Lecture 12

Today’s plan Moving from the bi-variate to the multivariate Looking at how the multivariate equation relates to the bi-variate equation Derivation The difference between true and estimated models Lecture 12

Introduction In multivariate regressions, your number of X variables are restricted by (n-k) > 0, where k is the number of parameters in your model In the bi-variate case we had If (n-k)  0 we wouldn’t be able to calculate test statistics We will use an example where earnings is our dependent variable, years of schooling (YRS_SCH) is X1, and age is X2 Lecture 12

Derivation The rules for the derivation of the parameters are the same as for the bi-variate world Our g function will be g(a, b1, b2) We will still want to minimize  e2 Our model will be Y = a + b1X1 + b2X2 + e We can rewrite this in terms from deviations from mean values (coded variables): y = b1x1 + b2x2 + e Lecture 12

Derivation (2) We can rearrange our model in terms of e: e = Y - a - b1X1 - b2X2 Differentiating with respect to each of the parameters gives us: Lecture 12

Derivation (3) To get our estimate of we use the FOC that the sum of the errors equal zero. We substitute in for e and solve: As we include more variables, we need more terms to calculate the intercept Calculating is more complicated Lecture 12

Derivation (4) We have the first order conditions for The multivariate case is much more complicated than the bi-variate case, but the pattern remains the same Denominator still considers the variation in X The numerator still considers the variation of X1, X2, and Y Lecture 12

Derivation (6) The multivariate case is much more complicated than the bi-variate case, but the pattern remains the same Denominator still considers the variation in X The numerator still considers the variation of X1, X2, and Y Lecture 12

Matrix of products & cross-products This will help us calculate b1 and b2, as well as other test statistics we’ll need The matrix of products and cross-products is symmetric Lecture 12

Example On L12.xls there is an example of a matrix of products and cross-products that we’re interested in. This spreadsheet also shows that LINEST can also accommodate a multivariate regression From the spreadsheet we know: Lecture 12

Example (2) We can then calculate: We can also calculate Lecture 12

Y = 4.53 + 0.097 X1,where X1 is years of schooling Example (3) So now we can ask: What was the effect of including age? Had we not included age, our bi-variate regression equation would be: Y = 4.53 + 0.097 X1,where X1 is years of schooling Including age, the multivariate regression equation is: Y = 4.135 + 0.057 X1 + 0.023 X2 By including age, we reduce the coefficient on education (X1) by nearly a half! Lecture 12

True & estimated models A true model can come from: 1) Economic theory an example of this is the Cobb-Douglas production function Y=ALK the form is provided by economic theory we want to test if  +  = 1 2) Ad-hoc variable inclusion The justification for the variables comes from economic theory, but we include variables on the basis of significance in statistical tests An example: the Phillips Curve Lecture 12

where X1 is still years of education Omitted Variable Bias Let’s go back to the returns to education example in L12.xls and examine Omitted Variable Bias: Let’s assume that the true model is: Y = a + b1X1 + b2X2 + e But what if we instead estimate the following model: Y = a + b1X1 + u where X1 is still years of education Lecture 12

True & estimated models (3) Reasons why we might not estimate the true model we might not be able to collect the necessary data we might simply forget to include other variables such as age in the regression Let’s rewrite our equations in terms of deviations from the mean: True model: y = b1x1 + b2x2 + e Estimated model: y = b1x1 + u Lecture 12

Omitted variable bias Our estimate of the slope coefficient for the bi-variate model will be: If we know the true model we can plug it into the above expression and take the expectation to get: Lecture 12

Omitted variable bias (2) We can multiply out the terms and simplify the expression: Recall that one of our CLR assumptions is E(x1 e) = 0, so This represents the omitted variable bias Lecture 12

Omitted variable bias example Returning to the L13.xls example, we have If we think that then: This leads to a biased estimate of Lecture 12

Recap / what’s to come We learned that deriving the multivariate regression equation is similar to deriving the bi-variate case We worked with a matrix of products and cross-products We looked at the difference between true and estimated regression models We learned to calculate the omitted variable bias In the next few lectures we’ll be doing some more with multivariate models and applications Lecture 12

Unnecessary Variables What happens if variables that are included in the estimated model, are not relevant under the ‘true’ model. Estimated model: y = b1x1 + b2x2 + e True model: y = b1x1 + u If variables are unnecessary, they will not count in the estimated model. How to detect that: t-ratio hypothesis tests/Joint hypothesis tests using the F-distribution. Helps to make models parsimonious. Lecture 12