The Probit Model Alexander Spermann University of Freiburg SS 2008.

Slides:



Advertisements
Similar presentations
Statistical Analysis SC504/HS927 Spring Term 2008
Advertisements

Lesson 10: Linear Regression and Correlation
Brief introduction on Logistic Regression
Binary Logistic Regression: One Dichotomous Independent Variable
The Art of Model Building and Statistical Tests. 2 Outline The art of model building Using Software output The t-statistic The likelihood ratio test The.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Inference for Regression
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
Ch11 Curve Fitting Dr. Deshi Ye
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
1 BINARY CHOICE MODELS: PROBIT ANALYSIS In the case of probit analysis, the sigmoid function F(Z) giving the probability is the cumulative standardized.
Nguyen Ngoc Anh Nguyen Ha Trang
1Prof. Dr. Rainer Stachuletz Limited Dependent Variables P(y = 1|x) = G(  0 + x  ) y* =  0 + x  + u, y = max(0,y*)
Chapter 13 Multiple Regression
Binary Response Lecture 22 Lecture 22.
PSY 307 – Statistics for the Behavioral Sciences
Chapter 10 Simple Regression.
In previous lecture, we highlighted 3 shortcomings of the LPM. The most serious one is the unboundedness problem, i.e., the LPM may make the nonsense predictions.
Chapter 12 Multiple Regression
SIMPLE LINEAR REGRESSION
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
So far, we have considered regression models with dummy variables of independent variables. In this lecture, we will study regression models whose dependent.
Chapter 11 Multiple Regression.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
In previous lecture, we dealt with the unboundedness problem of LPM using the logit model. In this lecture, we will consider another alternative, i.e.
Topic 3: Regression.
Correlation and Regression Analysis
Simple Linear Regression and Correlation
BINARY CHOICE MODELS: LOGIT ANALYSIS
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: binary choice logit models Original citation: Dougherty, C. (2012) EC220.
SIMPLE LINEAR REGRESSION
AM Recitation 2/10/11.
Introduction to Linear Regression and Correlation Analysis
Chapter 13: Inference in Regression
Linear Regression and Correlation
Correlation and Linear Regression
Hypothesis Testing in Linear Regression Analysis
1 BINARY CHOICE MODELS: PROBIT ANALYSIS In the case of probit analysis, the sigmoid function is the cumulative standardized normal distribution.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
Go to Table of Content Single Variable Regression Farrokh Alemi, Ph.D. Kashif Haqqi M.D.
AN INTRODUCTION TO LOGISTIC REGRESSION ENI SUMARMININGSIH, SSI, MM PROGRAM STUDI STATISTIKA JURUSAN MATEMATIKA UNIVERSITAS BRAWIJAYA.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Political Science 30: Political Inquiry. Linear Regression II: Making Sense of Regression Results Interpreting SPSS regression output Coefficients for.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Logistic Regression Analysis Gerrit Rooks
Qualitative and Limited Dependent Variable Models ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
04/19/2006Econ 6161 Econ 616 – Spring 2006 Qualitative Response Regression Models Presented by Yan Hu.
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Logit Models Alexander Spermann, University of Freiburg, SS Logit Models.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
BINARY LOGISTIC REGRESSION
Chapter 4 Basic Estimation Techniques
THE LOGIT AND PROBIT MODELS
Political Science 30: Political Inquiry
THE LOGIT AND PROBIT MODELS
Interval Estimation and Hypothesis Testing
Simple Linear Regression
Linear Regression and Correlation
Introductory Statistics
Presentation transcript:

The Probit Model Alexander Spermann University of Freiburg SS 2008

2 1. Notation and statistical foundations 2. Introduction to the Probit model 3. Application 4. Coefficients and marginal effects 5. Goodness-of-fit 6. Hypothesis tests Course outline

3 Notation and statistical foundations

4 Notation and statistical foundations – Vectors  Column vector:  Transposed (row vector):  Inner product:

5  PDF: probability density function f(x)  Example: Normal distribution:  Example: Standard normal distribution: N(0,1), µ = 0, σ = 1 Notation and statistical foundations – density function

6  Standard logistic distribution: Exponential distribution:  Poisson distribution: Notation and statistical foundations – distibutions

7  CDF: cumulative distribution function F(x)  Example: Standard normal distribution:  The cdf is the integral of the pdf. Notation and statistical foundations – CDF

8  Rule I:  Rule II:  Rule III: Notation and statistical foundations – logarithms

9  Why not use OLS instead?  Nonlinear estimation, for example by maximum likelihood. Introduction to the Probit model – binary variables 1 0 OLS (linear) x x x x x x x x x

10  Latent variable: Unobservable variable y* which can take all values in (-∞, +∞).  Example: y* = Utility(Labour income) - Utility(Non labour income)  Underlying latent model: Introduction to the Probit model – latent variables

11  Probit is based on a latent model: Assumption: Error terms are independent and normally distributed: Introduction to the Probit model – latent variables because of symmetry

12  Example: Introduction to the Probit model – CDF 1-0,2=0,8 0 0,2 0,5 0,8 1

13  F(z) lies between zero and one  CDF of Probit: CDF of Logit: Introduction to the Probit model – CDF Probit vs. Logit

14  PDF of Probit: PDF of Logit: Introduction to the Probit model – PDF Probit vs. Logit

15  Joint density:  Log likelihood function: Introduction to the Probit model – The ML principle

16  The principle of ML: Which value of β maximizes the probability of observing the given sample? Introduction to the Probit model – The ML principle

17  Example taken from Greene, Econometric Analysis, 5. ed. 2003, ch  10 observations of a discrete distribution  Random sample: 5, 0, 1, 1, 0, 3, 2, 3, 4, 1  PDF:  Joint density :  Which value of θ makes occurance of the observed sample most probable? Introduction to the Probit model – Example

18 2 Introduction to the Probit model – Example

19 Application  Analysis of the effect of a new teaching method in economic sciences  Data: Source: Spector, L. and M. Mazzeo, Probit Analysis and Economic Education. In: Journal of Economic Education, 11, 1980, pp.37-44

20 Application – Variables  Grade Dependent variable. Indicates whether a student improved his grades after the new teaching method PSI had been introduced (0 = no, 1 = yes).  PSI Indicates if a student attended courses that used the new method (0 = no, 1 = yes).  GPA Average grade of the student  TUCE Score of an intermediate test which shows previous knowledge of a topic.

21 Application – Estimation  Estimation results of the model (output from Stata):

22 Application – Discussion  ML estimator: Parameters were obtained by maximization of the log likelihood function. Here: 5 iterations were necessary to find the maximum of the log likelihood function ( )  Interpretation of the estimated coefficients:  Estimated coefficients do not quantify the influence of the rhs variables on the probability that the lhs variable takes on the value one.  Estimated coefficients are parameters of the latent model.

23 Coefficients and marginal effects  The marginal effect of a rhs variable is the effect of an unit change of this variable on the probability P(Y = 1|X = x), given that all other rhs variables are constant:  Recap: The slope parameter of the linear regression model measures directly the marginal effect of the rhs variable on the lhs variable.

24 Coefficients and marginal effects  The marginal effect depends on the value of the rhs variable.  Therefore, there exists an individual marginal effect for each person of the sample:

25 Coefficients and marginal effects – Computation  Two different types of marginal effects can be calculated:  Average marginal effect Stata command: margin  Marginal effect at the mean: Stata command: mfx compute

26 Coefficients and marginal effects – Computation  Principle of the computation of the average marginal effects:  Average of individual marginal effects

27 Coefficients and marginal effects – Computation  Computation of average marginal effects depends on type of rhs variable:  Continuous variables like TUCE and GPA:  Dummy variable like PSI:

28 Coefficients and marginal effects – Interpretation  Interpretation of average marginal effects:  Continuous variables like TUCE and GPA: An infinitesimal change of TUCE or GPA changes the probability that the lhs variable takes the value one by X%.  Dummy variable like PSI: A change of PSI from zero to one changes the probability that the lhs variable takes the value one by X%.

29 Coefficients and marginal effects – Interpretation VariableEstimated marginal effectInterpretation GPA0.364If the average grade of a student goes up by an infinitesimal amount, the probability for the variable grade taking the value one rises by 36.4%. TUCE0.011Analog to GPA,with an increase of 1.1%. PSI0.374If the dummy variable changes from zero to one, the probability for the variable grade taking the value one rises by 37.4%.

30 Coefficients and marginal effects – Significance  Significance of a coefficient: test of the hypothesis whether a parameter is significantly different from zero.  The decision problem is similar to the t-test, wheras the probit test statistic follows a standard normal distribution. The z-value is equal to the estimated parameter divided by its standard error.  Stata computes a p-value which shows directly the significance of a parameter: z-valuep-valueInterpretation GPA : significant TUCE: 0,62 0,533insignificant PSI: 2,67 0,008significant

31 Coefficients and marginal effects  Only the average of the marginal effects is displayed.  The individual marginal effects show large variation: Stata command: margin, table

32 Coefficients and marginal effects  Variation of marginal effects may be quantified by the confidence intervals of the marginal effects.  In which range one can expect a coefficient of the population?  In our example: Estimated coefficient Confidence interval (95%) GPA: 0, , ,782 TUCE: 0, , ,025 PSI: 0,374 0, ,626

33 Coefficients and marginal effects  What is calculated by mfx ?  Estimation of the marginal effect at the sample mean. Sample mean

34 Goodness of fit  Goodness of fit may be judged by McFaddens Pseudo R².  Measure for proximity of the model to the observed data.  Comparison of the estimated model with a model which only contains a constant as rhs variable.  : Likelihood of model of interest.  : Likelihood with all coefficients except that of the intercept restricted to zero.  It always holds that

35 Goodness of fit  The Pseudo R² is defined as:  Similar to the R² of the linear regression model, it holds that  An increasing Pseudo R² may indicate a better fit of the model, whereas no simple interpretation like for the R² of the linear regression model is possible.

36 Goodness of fit  A high value of R² McF does not necessarily indicate a good fit, however, as R² McF = 1 if = 0.  R² McF increases with additional rhs variables. Therefore, an adjusted measure may be appropriate:  Further goodness of fit measures: R² of McKelvey and Zavoinas, Akaike Information Criterion (AIC), etc. See also the Stata command fitstat.

37 Hypothesis tests  Likelihood ratio test: possibility for hypothesis testing, for example for variable relevance.  Basic principle: Comparison of the log likelihood functions of the unrestricted model (ln L U ) and that of the restricted model (ln L R )  Test statistic:  The test statistic follows a χ² distribution with degrees of freedom equal to the number of restrictions. ̃̃

38 Hypothesis tests  Null hypothesis: All coefficients except that of the intercept are equal to zero.  In the example:  Prob > chi2 =  Interpretation: The hypothesis that all coefficients are equal to zero can be rejected at the 1 percent significance level.