Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: testing a hypothesis relating to a regression coefficient (2010/2011.

Slides:



Advertisements
Similar presentations
CHOW TEST AND DUMMY VARIABLE GROUP TEST
Advertisements

EC220 - Introduction to econometrics (chapter 5)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: slope dummy variables Original citation: Dougherty, C. (2012) EC220 -
Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: a Monte Carlo experiment Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: introduction to maximum likelihood estimation Original citation: Dougherty,
Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: testing a hypothesis relating to a regression coefficient Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: interactive explanatory variables Original citation: Dougherty, C. (2012)
HETEROSCEDASTICITY-CONSISTENT STANDARD ERRORS 1 Heteroscedasticity causes OLS standard errors to be biased is finite samples. However it can be demonstrated.
EC220 - Introduction to econometrics (chapter 7)
EC220 - Introduction to econometrics (chapter 2)
00  sd  0 –sd  0 –1.96sd  0 +sd 2.5% CONFIDENCE INTERVALS probability density function of X null hypothesis H 0 :  =  0 In the sequence.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: expected value of a function of a random variable Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 6) Slideshow: variable misspecification iii: consequences for diagnostics Original.
TESTING A HYPOTHESIS RELATING TO THE POPULATION MEAN 1 This sequence describes the testing of a hypothesis at the 5% and 1% significance levels. It also.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: confidence intervals Original citation: Dougherty, C. (2012) EC220.
EC220 - Introduction to econometrics (chapter 1)
1 INTERPRETATION OF A REGRESSION EQUATION The scatter diagram shows hourly earnings in 2002 plotted against years of schooling, defined as highest grade.
EC220 - Introduction to econometrics (review chapter)
TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT This sequence describes the testing of a hypotheses relating to regression coefficients. It is.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: prediction Original citation: Dougherty, C. (2012) EC220 - Introduction.
SLOPE DUMMY VARIABLES 1 The scatter diagram shows the data for the 74 schools in Shanghai and the cost functions derived from a regression of COST on N.
1 In the previous sequence, we were performing what are described as two-sided t tests. These are appropriate when we have no information about the alternative.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: precision of the multiple regression coefficients Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: semilogarithmic models Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: nonlinear regression Original citation: Dougherty, C. (2012) EC220 -
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: maximum likelihood estimation of regression coefficients Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: Chow test Original citation: Dougherty, C. (2012) EC220 - Introduction.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: the normal distribution Original citation: Dougherty, C. (2012)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy variable classification with two categories Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: two sets of dummy variables Original citation: Dougherty, C. (2012) EC220.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
EC220 - Introduction to econometrics (review chapter)
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: sampling and estimators Original citation: Dougherty, C. (2012)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: the effects of changing the reference category Original citation: Dougherty,
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy classification with more than two categories Original citation:
DUMMY CLASSIFICATION WITH MORE THAN TWO CATEGORIES This sequence explains how to extend the dummy variable technique to handle a qualitative explanatory.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: Tobit models Original citation: Dougherty, C. (2012) EC220 - Introduction.
1 INTERACTIVE EXPLANATORY VARIABLES The model shown above is linear in parameters and it may be fitted using straightforward OLS, provided that the regression.
1 TWO SETS OF DUMMY VARIABLES The explanatory variables in a regression model may include multiple sets of dummy variables. This sequence provides an example.
Confidence intervals were treated at length in the Review chapter and their application to regression analysis presents no problems. We will not repeat.
1 t TEST OF A HYPOTHESIS RELATING TO A POPULATION MEAN The diagram summarizes the procedure for performing a 5% significance test on the slope coefficient.
1 PROXY VARIABLES Suppose that a variable Y is hypothesized to depend on a set of explanatory variables X 2,..., X k as shown above, and suppose that for.
F TEST OF GOODNESS OF FIT FOR THE WHOLE EQUATION 1 This sequence describes two F tests of goodness of fit in a multiple regression model. The first relates.
MULTIPLE REGRESSION WITH TWO EXPLANATORY VARIABLES: EXAMPLE 1 This sequence provides a geometrical interpretation of a multiple regression model with two.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 12) Slideshow: footnote: the Cochrane-Orcutt iterative process Original citation: Dougherty,
TYPE II ERROR AND THE POWER OF A TEST A Type I error occurs when the null hypothesis is rejected when it is in fact true. A Type II error occurs when the.
Simple regression model: Y =  1 +  2 X + u 1 We have seen that the regression coefficients b 1 and b 2 are random variables. They provide point estimates.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 9) Slideshow: instrumental variable estimation: variation Original citation: Dougherty,
. reg LGEARN S WEIGHT85 Source | SS df MS Number of obs = F( 2, 537) = Model |
Christopher Dougherty EC220 - Introduction to econometrics (chapter 6) Slideshow: multiple restrictions and zero restrictions Original citation: Dougherty,
COST 11 DUMMY VARIABLE CLASSIFICATION WITH TWO CATEGORIES 1 This sequence explains how you can include qualitative explanatory variables in your regression.
1 11 00 This sequence explains the logic behind a one-sided t test. ONE-SIDED t TESTS  0 +sd  0 –sd  0 –2sd  1 +2sd 2.5% null hypothesis:H 0 :
RAMSEY’S RESET TEST OF FUNCTIONAL MISSPECIFICATION 1 Ramsey’s RESET test of functional misspecification is intended to provide a simple indicator of evidence.
1 NONLINEAR REGRESSION Suppose you believe that a variable Y depends on a variable X according to the relationship shown and you wish to obtain estimates.
1 CHANGES IN THE UNITS OF MEASUREMENT Suppose that the units of measurement of Y or X are changed. How will this affect the regression results? Intuitively,
SEMILOGARITHMIC MODELS 1 This sequence introduces the semilogarithmic model and shows how it may be applied to an earnings function. The dependent variable.
GRAPHING A RELATIONSHIP IN A MULTIPLE REGRESSION MODEL The output above shows the result of regressing EARNINGS, hourly earnings in dollars, on S, years.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: confidence intervals Original citation: Dougherty, C. (2012) EC220 -
1 REPARAMETERIZATION OF A MODEL AND t TEST OF A LINEAR RESTRICTION Linear restrictions can also be tested using a t test. This involves the reparameterization.
F TESTS RELATING TO GROUPS OF EXPLANATORY VARIABLES 1 We now come to more general F tests of goodness of fit. This is a test of the joint explanatory power.
WHITE TEST FOR HETEROSCEDASTICITY 1 The White test for heteroscedasticity looks for evidence of an association between the variance of the disturbance.
1 COMPARING LINEAR AND LOGARITHMIC SPECIFICATIONS When alternative specifications of a regression model have the same dependent variable, R 2 can be used.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 1) Slideshow: simple regression model Original citation: Dougherty, C. (2012) EC220.
VARIABLE MISSPECIFICATION I: OMISSION OF A RELEVANT VARIABLE In this sequence and the next we will investigate the consequences of misspecifying the regression.
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Presentation transcript:

Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: testing a hypothesis relating to a regression coefficient (2010/2011 version) Original citation: Dougherty, C. (2012) EC220 - Introduction to econometrics (chapter 2). [Teaching Resource] © 2012 The Author This version available at: Available in LSE Learning Resources Online: May 2012 This work is licensed under a Creative Commons Attribution-ShareAlike 3.0 License. This license allows the user to remix, tweak, and build upon the work even for commercial purposes, as long as the user credits the author and licenses their new creations under the identical terms

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT This sequence describes the testing of a hypotheses relating to regression coefficients. We start by summarizing the basic procedure for a t test discussed in the Review chapter. 1

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT For the theory behind the t test, see the Review chapter. We will summarize only the procedure here. 2

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT We have a random variable X with unknown population mean  and variance  2. We wish to test a null hypothesis H 0 :  =  0 against the alternative H 1 :  ≠  0, given a sample of n observations. 3

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT We calculate the sample mean X and its standard error, and compute the t statistic shown. We reject H 0 if the absolute value of the t statistic is greater than the critical value of t, given the chosen significance level for the test. 4

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT In the case of the regression model, we are concerned with testing hypotheses relating to the unknown parameters of the relationship. Here also we perform t tests. The theory in principle is exactly the same and the procedure is parallel in an obvious way. 5

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT The t statistic is the difference between the estimated coefficient and its hypothesized value, divided by the standard error of the coefficient. We reject the null hypothesis if the absolute value is greater than the critical value of t, given the chose significance level. 6

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT There is one important difference. When locating the critical value of t, one must take account of the number of degrees of freedom. In the case of the random variable X, this is n – 1, where n is the number of observations in the sample. 7

REVIEW CHAPTER A random variable X is distributed with unknown population mean  and variance  2. The null and alternative hypotheses are H 0 :  =  0, H 1 :  ≠  0. To test H 0, we calculate the sample mean X and its standard error, and compute the test statistic. We reject H 0 if for a chosen significance level. REGRESSION MODEL Assuming that the true relationship is and that, given a sample of data, we have fitted the model. The null and alternative hypotheses are To test H 0, we compute. We reject H 0 if for a chosen significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT In the case of the regression model, the number of degrees of freedom is n – k, where n is the number of observations in the sample and k is the number of parameters (  coefficients). For the simple regression model above, it is n – 2. 8

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT As an illustration, we will consider a model relating price inflation to wage inflation. p is the percentage annual rate of growth of prices and w is the percentage annual rate of growth of wages. Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT We will test the hypothesis that the rate of price inflation is equal to the rate of wage inflation. The null hypothesis is therefore H 0 :  2 = 1.0. (We should also test  1 = 0.) Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Suppose that the regression result is as shown (standard errors in parentheses). Our actual estimate of the slope coefficient is only We will check whether we should reject the null hypothesis. Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT We compute the t statistic by subtracting the hypothetical true value from the sample estimate and dividing by the standard error. It comes to –1.80. Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT There are 20 observations in the sample. We have estimated 2 parameters, so there are 18 degrees of freedom. Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT The critical value of t with 18 degrees of freedom is at the 5% level. The absolute value of the t statistic is less than this, so we do not reject the null hypothesis. Two-sided t tests

15 In practice it is unusual to have a feeling for the actual value of the coefficients. Very often the objective of the analysis is to demonstrate that Y is influenced by X, without having any specific prior notion of the actual coefficients of the relationship. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

16 In this case it is usual to define  2 = 0 as the null hypothesis. In words, the null hypothesis is that X does not influence Y. We then try to demonstrate that the null hypothesis is false. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

17 For the null hypothesis  2 = 0, the t statistic reduces to the estimate of the coefficient divided by its standard error. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

18 This ratio is commonly called the t statistic for the coefficient and it is automatically printed out as part of the regression results. To perform the test for a given significance level, we compare the t statistic directly with the critical value of t for that significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | Here is the output from the earnings function fitted in a previous slideshow, with the t statistics highlighted. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | You can see that the t statistic for the coefficient of S is enormous. We would reject the null hypothesis that schooling does not affect earnings at the 1% significance level (critical value about 2.59). TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | In this case we could go further and reject the null hypothesis that schooling does not affect earnings at the 0.1% significance level. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | The advantage of reporting rejection at the 0.1% level, instead of the 1% level, is that the risk of mistakenly rejecting the null hypothesis of no effect is now only 0.1% instead of 1%. The result is therefore even more convincing. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | The t statistic for the intercept is also enormous. However, since the intercept does not hve any meaning, it does not make sense to perform a t test on it. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | The next column in the output gives what are known as the p values for each coefficient. This is the probability of obtaining the corresponding t statistic as a matter of chance, if the null hypothesis H 0 :  = 0 is true. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | If you reject the null hypothesis H 0 :  = 0, this is the probability that you are making a mistake and making a Type I error. It therefore gives the significance level at which the null hypothesis would just be rejected. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | If p = 0.05, the null hypothesis could just be rejected at the 5% level. If it were 0.01, it could just be rejected at the 1% level. If it were 0.001, it could just be rejected at the 0.1% level. This is assuming that you are using two-sided tests. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | In the present case p = 0 to three decimal places for the coefficient of S. This means that we can reject the null hypothesis H 0 :  2 = 0 at the 0.1% level, without having to refer to the table of critical values of t. (Testing the intercept does not make sense in this regression.) TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | It is a more informative approach to reporting the results of test and widely used in the medical literature. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | However in economics standard practice is to report results referring to 5% and 1% significance levels, and sometimes to the 0.1% level (when one can reject at that level). TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Two-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ Returning to the price inflation/wage inflation model, we saw that we could not reject the null hypothesis  2 = 1, even at the 5% significance level. That was using a two-sided test. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 ≠ TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT However, in practice, improvements in productivity may cause the rate of cost inflation, and hence that of price inflation, to be lower than that of wage inflation. One-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 < TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Certainly, improvements in productivity will not cause price inflation to be greater than wage inflation and so in this case we are justified in ruling out  2 > 1. We are left with H 0 :  2 = 1 and H 1 :  2 < 1. One-sided t tests

Example: p =  1 +  2 w + u Null hypothesis: H 0 :  2 = 1.0 Alternative hypothesis: H 1 :  2 < TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Thus we can perform a one-sided test, for which the critical value of t with 18 degrees of freedom at the 5% significance level is Now we can reject the null hypothesis and conclude that price inflation is significantly lower than wage inflation, at the 5% significance level. One-sided t tests

34 Now we will consider the special, but very common, case H 0 :  2 = 0. TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 One-sided t tests

35 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 It occurs when you wish to demonstrate that a variable X influences another variable Y. You set up the null hypothesis that X has no effect (  2 = 0) and try to reject H 0. One-sided t tests

36 probability density function of b 2 0 The figure shows the distribution of b 2, conditional on H 0 :  2 = 0 being true. For simplicity, we initially assume that we know the standard deviation. 2.5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 = 0 reject H 0 do not reject H sd–1.96 sd TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

37 probability density function of b 2 0 If you use a two-sided 5% significance test, your estimate must be 1.96 standard deviations above or below 0 if you are to reject H % null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 = 0 reject H 0 do not reject H sd–1.96 sd TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

38 probability density function of b 2 0 However, if you can justify the use of a one-sided test, for example with H 0 :  2 > 0, your estimate has to be only 1.65 standard deviations above 0. reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 This makes it easier to reject H 0 and thereby demonstrate that Y really is influenced by X (assuming that your model is correctly specified). 39 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 Suppose that Y is genuinely determined by X and that the true (unknown) coefficient is  2, as shown. 40 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 Suppose that we have a sample of observations and calculate the estimated slope coefficient, b 2. If it is as shown in the diagram, what do we conclude when we test H 0 ? 41 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 The answer is that b 2 lies in the rejection region. It makes no difference whether we perform a two-sided test or a one-sided test. We come to the correct conclusion. 42 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 What do we conclude if b 2 is as shown? We fail to reject H 0, irrespective of whether we perform a two-sided test or a two-dsided test. We would make a Type II error in either case. 43 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 What do we conclude if b 2 is as shown here? In the case of a two-sided test, b 2 is not in the rejection region. We are unable to reject H TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 This means that we are unable to demonstrate that X has a significant effect on Y. This is disappointing, because we were hoping to demonstrate that X is a determinant of Y. 45 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 However, if we are in a position to perform a one-sided test, b 2 does lie in the rejection region and so we have demonstrated that X has a significant effect on Y (at the 5% significance level, of course). 46 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 Thus we get a positive finding that we could not get with a two-sided test. 47 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 22 b2b2

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 To put this reasoning more formally, the power of a one-sided test is greater than that of a two-sided test. The blue area shows the probability of making a Type II error using a two- sided test. It is the area under the true curve to the left of the rejection region. 48 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 2 sd

probability density function of b 2 0 reject H 0 do not reject H 0 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 The red area shows the probability of making a Type II error using a one-sided test. It is smaller. Since the power of a test is (1 – probability of making a Type II error when H 0 is false), the power of a one-sided test is greater than that of a two-sided test. 49 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests 2 sd

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 In all of this, we have assumed that we knew the standard deviation of the distribution of b 2. In practice, of course, the standard deviation has to be estimated as the standard error, and the t distribution is the relevant distribution. However, the logic is exactly the same. 50 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 At any given significance level, the critical value of t for a one-sided test is lower than that for a two-sided test. 51 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

probability density function of b 2 0 reject H 0 do not reject H sd 5% null hypothesis:H 0 :  2 = 0 alternative hypothesis:H 1 :  2 > 0 Hence, if H 0 is false, the risk of not rejecting it, thereby making a Type II error, is smaller, and so the power of a one-sided test is greater than that of a two-sided test. 52 TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT One-sided t tests

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 Reject H 0 ifor Do not reject H 0 if Confidence intervals were treated at length in the Review chapter and their application to regression analysis presents no problems. We will not repeat the graphical explanation. We will just provide the mathematical derivation in the context of a regression. Confidence intervals 53

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 Reject H 0 ifor Do not reject H 0 if From the initial discussion in this section, we saw that, given the theoretical model Y =  1 +  2 X + u and a fitted model, the regression coefficient b 2 and the hypothetical value of  2 are incompatible if either of the inequalities shown is valid. Confidence intervals 54

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 Reject H 0 ifor Do not reject H 0 if Multiplying through by the standard error of b 2, the conditions for rejecting H 0 can be written as shown. Confidence intervals 55

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 Reject H 0 ifor Do not reject H 0 if The inequalities may then be re-arranged as shown. Confidence intervals 56

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Model: Y =  1 +  2 X + u Null hypothesis: H 0 :  2 = 0 Alternative hypothesis: H 1 :  2 ≠ 0 Reject H 0 ifor Do not reject H 0 if We can then obtain the confidence interval for  2, being the set of all values that would not be rejected, given the sample estimate b 2. To make it operational, we need to select a significance level and determine the corresponding critical value of t. Confidence intervals 57

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT For an example of the construction of a confidence interval, we will return to the earnings function fitted earlier. We will construct a 95% confidence interval for  2.. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | – x ≤  2 ≤ x ≤  2 ≤ Confidence intervals 58

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT The point estimate  2 is and its standard error is reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | – x ≤  2 ≤ x ≤  2 ≤ Confidence intervals 59

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT The critical value of t at the 5% significance level with 538 degrees of freedom is reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | – x ≤  2 ≤ x ≤  2 ≤ Confidence intervals 60

TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT Hence we establish that the confidence interval is from to Stata actually computes the 95% confidence interval as part of its default output, to The discrepancy in the lower limit is due to rounding error in the calculations we have made. Confidence intervals. reg EARNINGS S Source | SS df MS Number of obs = F( 1, 538) = Model | Prob > F = Residual | R-squared = Adj R-squared = Total | Root MSE = EARNINGS | Coef. Std. Err. t P>|t| [95% Conf. Interval] S | _cons | – x ≤  2 ≤ x ≤  2 ≤

Copyright Christopher Dougherty These slideshows may be downloaded by anyone, anywhere for personal use. Subject to respect for copyright and, where appropriate, attribution, they may be used as a resource for teaching an econometrics course. There is no need to refer to the author. The content of this slideshow comes from Section 2.6 of C. Dougherty, Introduction to Econometrics, fourth edition 2011, Oxford University Press. Additional (free) resources for both students and instructors may be downloaded from the OUP Online Resource Centre Individuals studying econometrics on their own and who feel that they might benefit from participation in a formal course should consider the London School of Economics summer school course EC212 Introduction to Econometrics or the University of London International Programmes distance learning course 20 Elements of Econometrics