Chengyuan Yin School of Mathematics

Slides:



Advertisements
Similar presentations
Regression Analysis.
Advertisements

Tests of Static Asset Pricing Models
The Multiple Regression Model.
Brief introduction on Logistic Regression
Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
3. Binary Choice – Inference. Hypothesis Testing in Binary Choice Models.
CHAPTER 8 MULTIPLE REGRESSION ANALYSIS: THE PROBLEM OF INFERENCE
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Part 8: Hypothesis Testing 8-1/50 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Objectives (BPS chapter 24)
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Hypothesis testing Some general concepts: Null hypothesisH 0 A statement we “wish” to refute Alternative hypotesisH 1 The whole or part of the complement.
Point estimation, interval estimation
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
T-test.
Chapter 11 Multiple Regression.
Horng-Chyi HorngStatistics II41 Inference on the Mean of a Population - Variance Known H 0 :  =  0 H 0 :  =  0 H 1 :    0, where  0 is a specified.
5-3 Inference on the Means of Two Populations, Variances Unknown
Part 7: Multiple Regression Analysis 7-1/54 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Inference for regression - Simple linear regression
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Part 9: Hypothesis Testing /29 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
May 2004 Prof. Himayatullah 1 Basic Econometrics Chapter 5: TWO-VARIABLE REGRESSION: Interval Estimation and Hypothesis Testing.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
6. Simple Regression and OLS Estimation Chapter 6 will expand on concepts introduced in Chapter 5 to cover the following: 1) Estimating parameters using.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Multiple Regression Analysis: Inference
Heteroscedasticity Chapter 8
Chapter 14 Introduction to Multiple Regression
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
6. Simple Regression and OLS Estimation
Chapter 4. Inference about Process Quality
Multiple Regression Analysis: Inference
THE LINEAR REGRESSION MODEL: AN OVERVIEW
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Chapter 2 Simple Comparative Experiments
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Hypothesis Tests for a Population Mean in Practice
Discrete Event Simulation - 4
Model Comparison: some basic concepts
Undergraduated Econometrics
Review of Statistical Inference
Statistical Process Control
Interval Estimation and Hypothesis Testing
Chengyuan Yin School of Mathematics
Simple Linear Regression
Tutorial 1: Misspecification
Econometrics Analysis
Heteroskedasticity.
Simple Linear Regression
Chapter 7: The Normality Assumption and Inference with OLS
Chengyuan Yin School of Mathematics
Inference on the Mean of a Population -Variance Known
Econometrics I Professor William Greene Stern School of Business
The Multiple Regression Model
Chapter 9 Dummy Variables Undergraduated Econometrics Page 1
Econometrics I Professor William Greene Stern School of Business
Presentation transcript:

Chengyuan Yin School of Mathematics Econometrics Chengyuan Yin School of Mathematics

8. Hypothesis Testing in the Linear Regression Model Econometrics 8. Hypothesis Testing in the Linear Regression Model

Classical Hypothesis Testing We are interested in using the linear regression to establish or cast doubt on the validity of a theory about the real world counterpart to our statistical model. The model is used to test hypotheses about the underlying data generating process.

Inference in the Linear Model Hypothesis testing: Formulating hypotheses: linear restrictions as a general framework Substantive restrictions: What is a "testable hypothesis?" Nested vs. nonnested models Methodological issues Classical (likelihood based approach): Are the data consistent with the hypothesis? Bayesian approach: How do the data affect our prior odds?  The posterior odds ratio.

Testing a Hypothesis About a Parameter: Confidence Interval bk = the point estimate Std.Dev[bk] = sqr{[σ2(X’X)-1]kk} = vk Assume normality of ε for now: bk ~ N[βk,vk2] for the true βk. (bk-βk)/vk ~ N[0,1] Consider a range of plausible values of βk given the point estimate bk. bk +/- sampling error. Measured in standard error units, |(bk – βk)/ vk| < z* Larger z*  greater probability (“confidence”) Given normality, e.g., z* = 1.96  95%, z*=1.64590% Plausible range for βk then is bk ± z* vk

Estimating the Confidence Interval Assume normality of ε for now: bk ~ N[βk,vk2] for the true βk. (bk-βk)/vk ~ N[0,1] vk = [σ2(X’X)-1]kk is not known because σ2 must be estimated. Using s2 instead of σ2, (bk-βk)/est.(vk) ~ t[n-K]. (Proof: ratio of normal to sqr(chi-squared)/df is pursued in your text.) Use critical values from t distribution instead of standard normal.

Testing a Hypothesis Using a Confidence Interval Given the range of plausible values The confidence interval approach. Testing the hypothesis that a coefficient equals zero or some other particular value: Is the hypothesized value in the confidence interval? Is the hypothesized value within the range of plausible values

Wald Distance Measure Testing more generally about a single parameter. Sample estimate is bk Hypothesized value is βk How far is βk from bk? If too far, the hypothesis is inconsistent with the sample evidence. Measure distance in standard error units t = (bk - βk)/Estimated vk. If t is “large” (larger than critical value), reject the hypothesis.

The Wald Statistic

Robust Tests The Wald test generally will (when properly constructed) be more robust to failures of the narrow model assumptions than the t or F Reason: Based on “robust” variance estimators and asymptotic results that hold in a wide range of circumstances. Analysis: Later in the course – after developing asymptotics.

The General Linear Hypothesis: H0: R - q = 0 A unifying departure point: Regardless of the hypothesis, least squares is unbiased. Two approaches (1) Is Rb - q close to 0? Basing the test on the discrepancy vector: m = Rb - q. Using the Wald criterion: m(Var[m])-1m has a chi-squared distribution with J degrees of freedom But, Var[m] = R[2(X’X)-1]R. If we use our estimate of 2, we get an F[J,n-K], instead. (Note, this is based on using ee/(n-K) to estimate 2.) (2) We know that imposing the restrictions leads to a loss of fit. R2 must go down. Does it go down “a lot?” (I.e., significantly?). R2 = unrestricted model, R*2 = restricted model fit. F = { (R2 - R*2)/J } / [(1 - R2)/(n-K)] = F[J,n-K]. These are the same in the linear model

t and F statistics An important relationship between t and F For a single restriction, m = r’b - q. The variance is r’(Var[b])r The distance measure is m / standard error of m. The t-ratio is the square root of the F ratio.

Lagrange Multiplier Statistics Specific to the classical model: Recall the Lagrange multipliers:  = [R(XX)-1R]-1m. Suppose we just test H0:  = 0, using the Wald criterion. The resulting test statistic is just JF where F is the F statistic above. This is to be taken as a chi-squared statistic. (Note, again, using ee/(n-K) to estimate 2. If ee/n, instead, the more formal, likelihood based statistic results.)

Application Time series regression, LogG = 1 + 2logY + 3logPG + 4logPNC + 5logPUC + 6logPPT + 7logPN + 8logPD + 9logPS +  Period = 1960 - 1995. A significant event occurs in October 1973. We will be interested to know if the model 1960 to 1973 is the same as from 1974 to 1995. Note that all coefficients in the model are elasticities.

Full Model

Test about one Parameter Is the price of public transportation really relevant? H0 : 6 = 0. Confidence interval: b6  t(.95,27)  Standard error = .1279  2.052(.0791) = .1279  .1623 = (-.0344 ,.2902) Contains 0.0. Do not reject hypothesis Distance measure: (b6 - 0) / sb6 = (.1279 - 0) / .0791 = 1.617 < 2.052. Regression fit if drop? Without LPPT, R-squared= .988215 Compare R2, was .989256, F(1,27) = [(.989256 - .988215)/1]/[(1-.989256)/(36-9)] = 2.616 = 1.61722 (!)

Sum of Coefficients

Imposing the Restrictions

Joint Hypotheses

Using the Wald Statistic

Details Which restriction is the problem? After building the restrictions into the model and computing restricted and unrestricted regressions: Based on R2s, F = ((.989256 - .997237)/2)/((1-.989256)/(36-9)) = -.10.02824832465 (!) What's wrong?

Chow Test

Algebra for the Chow Test