In previous lecture, we highlighted 3 shortcomings of the LPM. The most serious one is the unboundedness problem, i.e., the LPM may make the nonsense predictions.

Slides:



Advertisements
Similar presentations
Brief introduction on Logistic Regression
Advertisements

EC220 - Introduction to econometrics (chapter 10)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: slope dummy variables Original citation: Dougherty, C. (2012) EC220 -
Logit & Probit Regression
Introduction to Logistic Regression In Stata Maria T. Kaylen, Ph.D. Indiana Statistical Consulting Center WIM Spring 2014 April 11, 2014, 3:00-4:30pm.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: interactive explanatory variables Original citation: Dougherty, C. (2012)
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
Lecture 9 Today: Ch. 3: Multiple Regression Analysis Example with two independent variables Frisch-Waugh-Lovell theorem.
INTERPRETATION OF A REGRESSION EQUATION
1 BINARY CHOICE MODELS: PROBIT ANALYSIS In the case of probit analysis, the sigmoid function F(Z) giving the probability is the cumulative standardized.
Multilevel Models 4 Sociology 8811, Class 26 Copyright © 2007 by Evan Schofer Do not copy or distribute without permission.
Logistic Regression Multivariate Analysis. What is a log and an exponent? Log is the power to which a base of 10 must be raised to produce a given number.
QUALITATIVE AND LIMITED DEPENDENT VARIABLE MODELS.
SLIDE 1IS 240 – Spring 2010 Logistic Regression The logistic function: The logistic function is useful because it can take as an input any.
An Introduction to Logistic Regression JohnWhitehead Department of Economics Appalachian State University.
So far, we have considered regression models with dummy variables of independent variables. In this lecture, we will study regression models whose dependent.
In previous lecture, we dealt with the unboundedness problem of LPM using the logit model. In this lecture, we will consider another alternative, i.e.
© Christopher Dougherty 1999–2006 VARIABLE MISSPECIFICATION I: OMISSION OF A RELEVANT VARIABLE We will now investigate the consequences of misspecifying.
TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT This sequence describes the testing of a hypotheses relating to regression coefficients. It is.
Chapter 4 – Nonlinear Models and Transformations of Variables.
SLOPE DUMMY VARIABLES 1 The scatter diagram shows the data for the 74 schools in Shanghai and the cost functions derived from a regression of COST on N.
BINARY CHOICE MODELS: LOGIT ANALYSIS
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: semilogarithmic models Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: nonlinear regression Original citation: Dougherty, C. (2012) EC220 -
TOBIT ANALYSIS Sometimes the dependent variable in a regression model is subject to a lower limit or an upper limit, or both. Suppose that in the absence.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy variable classification with two categories Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: two sets of dummy variables Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: the effects of changing the reference category Original citation: Dougherty,
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy classification with more than two categories Original citation:
DUMMY CLASSIFICATION WITH MORE THAN TWO CATEGORIES This sequence explains how to extend the dummy variable technique to handle a qualitative explanatory.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: Tobit models Original citation: Dougherty, C. (2012) EC220 - Introduction.
1 INTERACTIVE EXPLANATORY VARIABLES The model shown above is linear in parameters and it may be fitted using straightforward OLS, provided that the regression.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: binary choice logit models Original citation: Dougherty, C. (2012) EC220.
1 TWO SETS OF DUMMY VARIABLES The explanatory variables in a regression model may include multiple sets of dummy variables. This sequence provides an example.
1 PROXY VARIABLES Suppose that a variable Y is hypothesized to depend on a set of explanatory variables X 2,..., X k as shown above, and suppose that for.
Methods Workshop (3/10/07) Topic: Event Count Models.
1 BINARY CHOICE MODELS: PROBIT ANALYSIS In the case of probit analysis, the sigmoid function is the cumulative standardized normal distribution.
F TEST OF GOODNESS OF FIT FOR THE WHOLE EQUATION 1 This sequence describes two F tests of goodness of fit in a multiple regression model. The first relates.
LOGISTIC REGRESSION A statistical procedure to relate the probability of an event to explanatory variables Used in epidemiology to describe and evaluate.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: exercise 5.2 Original citation: Dougherty, C. (2012) EC220 - Introduction.
Chapter 5: Dummy Variables. DUMMY VARIABLE CLASSIFICATION WITH TWO CATEGORIES 1 We’ll now examine how you can include qualitative explanatory variables.
(1)Combine the correlated variables. 1 In this sequence, we look at four possible indirect methods for alleviating a problem of multicollinearity. POSSIBLE.
Warsaw Summer School 2015, OSU Study Abroad Program Advanced Topics: Interaction Logistic Regression.
COST 11 DUMMY VARIABLE CLASSIFICATION WITH TWO CATEGORIES 1 This sequence explains how you can include qualitative explanatory variables in your regression.
The dangers of an immediate use of model based methods The chronic bronchitis study: bronc: 0= no 1=yes poll: pollution level cig: cigarettes smokes per.
Logistic Regression Analysis Gerrit Rooks
1 NONLINEAR REGRESSION Suppose you believe that a variable Y depends on a variable X according to the relationship shown and you wish to obtain estimates.
SEMILOGARITHMIC MODELS 1 This sequence introduces the semilogarithmic model and shows how it may be applied to an earnings function. The dependent variable.
1 BINARY CHOICE MODELS: LINEAR PROBABILITY MODEL Economists are often interested in the factors behind the decision-making of individuals or enterprises,
1 REPARAMETERIZATION OF A MODEL AND t TEST OF A LINEAR RESTRICTION Linear restrictions can also be tested using a t test. This involves the reparameterization.
F TESTS RELATING TO GROUPS OF EXPLANATORY VARIABLES 1 We now come to more general F tests of goodness of fit. This is a test of the joint explanatory power.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
1 COMPARING LINEAR AND LOGARITHMIC SPECIFICATIONS When alternative specifications of a regression model have the same dependent variable, R 2 can be used.
The Probit Model Alexander Spermann University of Freiburg SS 2008.
VARIABLE MISSPECIFICATION I: OMISSION OF A RELEVANT VARIABLE In this sequence and the next we will investigate the consequences of misspecifying the regression.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Nonparametric Statistics
Lecture 18 Matched Case Control Studies
THE LOGIT AND PROBIT MODELS
Introduction to Logistic Regression
THE LOGIT AND PROBIT MODELS
Nonparametric Statistics
Logistic Regression.
Introduction to Logistic Regression
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Presentation transcript:

In previous lecture, we highlighted 3 shortcomings of the LPM. The most serious one is the unboundedness problem, i.e., the LPM may make the nonsense predictions that an event will occur with probability greater than 1 or less than 0. In this lecture, we will deal with this problem using the logit model. 1 Adapted from “Introduction to Econometrics” by Christopher Dougherty

The usual way of avoiding this problem is to hypothesize that the probability is a sigmoid (S-shaped) function of Z, F(Z), where Z is a function of the explanatory variables. The "input" is Z and the "output" is F(Z). BINARY CHOICE MODELS: LOGIT ANALYSIS 2

Several mathematical functions are sigmoid in character. One is the logistic function. The logistic function is useful because it can take as an input any value from negative infinity to positive infinity, whereas the output is confined to values between 0 and 1. The variable Z represents the exposure to some set of independent variables, in this case, Z =  1 +  2 X. Z is also known as logit. F(Z) represents the probability of an event occurring (probability of success, p), given that set of explanatory variables. 3

As Z goes to + , e –Z goes to 0 and p goes to 1 (but cannot exceed 1). As Z goes to - , e –Z goes to  and p goes to 0 (but cannot be below 0). 4

For a nonlinear model of this kind, maximum likelihood estimation is much superior to the use of OLS for estimating the parameters. ILLUSTRATION 1 Why do some people graduate from high school while others drop out? Our LPM model is in the form (see previous lecture): GRAD =  1 +  2 ASVABC +  i Our logit model is in the form: Z =  1 +  2 ASVABC i 5

6

. logit GRAD ASVABC Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Iteration 4: log likelihood = Logit estimates Number of obs = 540 LR chi2(1) = Prob > chi2 = Log likelihood = Pseudo R2 = GRAD | Coef. Std. Err. z P>|z| [95% Conf. Interval] ASVABC | _cons |

8

What is the probability that an individual with a score of 30 will graduate? When ASVABC = 30, An individual with a score of 30 has a 67 percent probability of graduating. 9

The coefficients of the Z function do not have any direct intuitive interpretation. However, we can use them to quantify the marginal effect of a change in ASVABC on the probability of graduating. We will do this theoretically for the general case where Z is a function of several explanatory variables. Since p is a function of Z, and Z is a function of the X variables, the marginal effect of X i on p can be written as: 10

(1) 11

(2) (3) Hence we obtain an expression for the marginal effect of X i on p. The marginal effect is not constant because it depends on the value of Z, which in turn depends on the values of the explanatory variables. A common procedure is to evaluate it for the sample means of the explanatory variables. 12

The sample mean of ASVABC in this sample is sum GRAD ASVABC Variable | Obs Mean Std. Dev. Min Max GRAD | ASVABC | When evaluated at the mean, Z is equal to

The marginal effect, evaluated at the mean, is therefore This implies that a one point increase in ASVABC would increase the probability of graduating from high school by 0.4%. 14

As noted earlier, the marginal effect does vary because it depends on the value of Z, which in turn depends on the values of X. How much is the marginal effect when ASVABC equal to 30? A one point increase in ASVABC then increases the probability of graduating by 2.9% 15

16

. logit GRAD ASVABC SM SF MALE Iteration 0: log likelihood = Iteration 1: log likelihood = Iteration 2: log likelihood = Iteration 3: log likelihood = Iteration 4: log likelihood = Iteration 5: log likelihood = Here we consider a multivariate model with a somewhat better specification. ILLUSTRATION 2 Why do some people graduate from high school while others drop out? 17

Logit estimates Number of obs = 540 LR chi2(4) = Prob > chi2 = Log likelihood = Pseudo R2 = GRAD | Coef. Std. Err. z P>|z| [95% Conf. Interval] ASVABC | SM | SF | MALE | _cons | sum GRAD ASVABC SM SF MALE Variable | Obs Mean Std. Dev. Min Max GRAD | ASVABC | SM | SF | MALE |

ASVABC SM11.58–0.023–0.269 SF MALE Constant1.00–3.252–3.252 Total3.514 We will estimate the marginal effects, putting all the explanatory variables equal to their sample means. Step 1: Calculate Z, when the X variables are equal to their sample means. 19

Step 2: Calculate Step 3: Calculate Note that: 20

We see that the effect of ASVABC is about the same as before. Mother's schooling (SM) has negligible effect and father's schooling (SF) has no discernible effect at all. Males have 0.4 percent higher probability of graduating than females. ASVABC SM11.58–0.023– –0.001 SF MALE

Suppose that we want to examine the effect of a teaching method known as PSI on the performance of students in GD The question was whether students exposed to the method scored higher on exams in the class. Data were collected from 32 students in two classes, one in which PSI was used and another in which a traditional teaching method was employed. 1) GPA — Grade point average before taking the class. 2) TUCE — the score on an exam given at the beginning of the term to test entering knowledge of the material. 3) PSI — a dummy variable indicating the teaching method used (1 = used Psi, 0 = other method). 4) GRADE — coded 1 if the final grade was an A, 0 if the final grade was a B or C. GRADE was the dependent variable, and of particular interest was whether PSI had a significant effect on GRADE. TUCE and GPA are included as control variables. ILLUSTRATION 3 22

. logit GRADE GPA TUCE PSI Logit estimates Number of obs = 32 LR chi2(4) = Prob > chi2 = Log likelihood = Pseudo R2 = GRADE | Coef. Std. Err. z P>|z| [95% Conf. Interval] GPA | TUCE | PSI | _cons |

GPA TUCE PSI Constant1.00–13.021– Total What is the probability of getting an A for a student who has a GPA of 3.0, is taught by traditional methods, and has a score of 20 on TUCE? This student has a 6.66% chance of getting an A. 24

GPA TUCE PSI Constant1.00–13.021– Total What is the probability of getting an A for the same student if he is placed in the PSI class? Hence, getting into the PSI class would substantially increase the chances of getting an A, i.e., the student would have about a 37% better chance. 25