QUALITATIVE AND LIMITED DEPENDENT VARIABLE MODELS.

Slides:



Advertisements
Similar presentations
Dummy Dependent variable Models
Advertisements

Econometric Modeling Through EViews and EXCEL
Brief introduction on Logistic Regression
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
© Copyright 2001, Alan Marshall1 Regression Analysis Time Series Analysis.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
Nguyen Ngoc Anh Nguyen Ha Trang
Models with Discrete Dependent Variables
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
Chapter 12 Simple Regression
Regression with a Binary Dependent Variable. Introduction What determines whether a teenager takes up smoking? What determines if a job applicant is successful.
The Simple Regression Model
An Introduction to Logistic Regression JohnWhitehead Department of Economics Appalachian State University.
So far, we have considered regression models with dummy variables of independent variables. In this lecture, we will study regression models whose dependent.
Chapter Topics Types of Regression Models
Qualitative and Limited Dependent Variable Models Prepared by Vera Tabakova, East Carolina University.
Topic 3: Regression.
An Introduction to Logistic Regression
Pertemua 19 Regresi Linier
Correlation and Regression Analysis
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Introduction to Regression Analysis, Chapter 13,
Chapter 12 Section 1 Inference for Linear Regression.
Simple Linear Regression Analysis
MODELS OF QUALITATIVE CHOICE by Bambang Juanda.  Models in which the dependent variable involves two ore more qualitative choices.  Valuable for the.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Regression and Correlation Methods Judy Zhong Ph.D.
Correlation and Linear Regression
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Qualitative and Limited Dependent Variable Models Adapted from Vera Tabakova’s notes ECON 4551 Econometrics II Memorial University of Newfoundland.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 7 Logistic Regression I.
AN INTRODUCTION TO LOGISTIC REGRESSION ENI SUMARMININGSIH, SSI, MM PROGRAM STUDI STATISTIKA JURUSAN MATEMATIKA UNIVERSITAS BRAWIJAYA.
Managerial Economics Demand Estimation & Forecasting.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Discussion of time series and panel models
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
7.4 DV’s and Groups Often it is desirous to know if two different groups follow the same or different regression functions -One way to test this is to.
Regression Analysis: Part 2 Inference Dummies / Interactions Multicollinearity / Heteroscedasticity Residual Analysis / Outliers.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice- Hall, Inc. Chap 14-1 Business Statistics: A Decision-Making Approach 6 th Edition.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Lynn Smith.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Multiple Regression Learning Objectives n Explain the Linear Multiple Regression Model n Interpret Linear Multiple Regression Computer Output n Test.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
Regression Analysis: A statistical procedure used to find relations among a set of variables B. Klinkenberg G
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
Lecturer: Ing. Martina Hanová, PhD. Business Modeling.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
PO 141: INTRODUCTION TO PUBLIC POLICY Summer I (2015) Claire Leavitt Boston University.
Chapter 4: Basic Estimation Techniques
Chapter 14 Introduction to Multiple Regression
BINARY LOGISTIC REGRESSION
Chapter 4 Basic Estimation Techniques
Limited Dependent Variables
Basic Estimation Techniques
THE LOGIT AND PROBIT MODELS
Multiple Regression Analysis and Model Building
Basic Estimation Techniques
Prepared by Lee Revere and John Large
Regression and Categorical Predictors
Presentation transcript:

QUALITATIVE AND LIMITED DEPENDENT VARIABLE MODELS

Model to describe choice behavior Dependent variables that are limited, that is the range of values is constrained Or The values are not completely observable

e.g. a worker decides to drive to work or not a high school graduate decides to go to college or not a household decides to purchase a house or rent why are some loan applications accepted and others not?

Qualitative Choice Models: Binary choice models Occurs when an individual is making a choice. Models with Binary Dependent Variables If the dependent variable assumes only 2 values 1 if the outcome is chosen and 0 if not

For these models, least squares estimation methods are not the best choices. In this case OLS is both biased and inconsistent OLS suffers from heteroscedasticity problem Instead, maximum likelihood estimation (MLE) is the usual method used

MLE of probit (or logit) discrete choice model Slopes can only be estimated up to a scale factor But coefficient signs and t-values have the usual interpretation

An example Y = Affairs = 1 if individual has had an affair X1 = Dummy for male X2 = Years of marriage X3 = Dummy for kids in the marriage X4 = Dummy for religion X5 = Years of education X6 = Dummy for Happy

Logit (MLE) estimate gives Y = X X X X X X6 X2, X4, X6 have p-values < 5% Logit coefficients do not directly measure marginal effects so its hard to interpret them. However, we can interpret the signs.

Odds ratio Marginal effects on the odds ratio. Suppose the coefficient on Happy under the odds ratio is 0.42, how do we interpret this number. Happy is a dummy variable  if a person switches from an unhappy relationship to a happy one, the odds ratio would be 42% of what it was before.

Suppose an individual had an odds ratio of 4. i.e. P(Y=1) = 4/5 and P(Y=0) = 1/5  there’s an 80% chance that an individual will have an affair. If the individual’s marriage becomes Happy, then the odds ratio becomes 42% higher as before.  4*0.42 = 1.68   There is a 63% chance the individual will have an affair

The linear probability model Suppose we wish to explain an individual’s choice between driving to work (private transportation) and taking the bus (public transportation) Individual’s choice can be represented by a dummy variable

Y = 1if individual drives to work = 0if takes bus We can collect a random sample of workers then the outcome Y will have a probability function where p = probability that y=1. E(y) = p

Explanatory variable = X Difference between time by bus and time by car Expectation: As x increases, an individual would be more inclined to drive to work  We expect a positive relationship between x and p (the probability to drive to work)

The linear regression model = the linear probability model = Y= E(y) + e Y = p + e Y = Bo + B1X + e This model is heteroscedastic, because the variance of the error term varies from one observation to another.

If we use OLS to estimate the model, we’d obtain

Consider what happens when you use this model to predict behavior: If you substitute different values of x in the the equation, you might obtain values of phat that are 1.Less than 0 (negative) 2.Greater than 1 Values that do not make sense as probabilities

Generally, in a linear regression model the slope coefficient (if positive) suggests that the increase in x will have a constant effect on y But in probability models (binary dependent variable model), the constant rate of increase is impossible because 0≥ p ≤1

To overcome this problem we use nonlinear probit and logit models. In E-views, choose Objects, New Object, Equation and select the Binary Estimation option. Specify your equation in the equation box. E-views uses Maximum Likelihood estimator

The estimates produces by a probit or logit model that look like slopes and intercepts in the output are actually standardized slopes and intercepts, known only up to a scale factor So the size of these coefficients is not viewed in the same way (as the change in Y for a one-unit change in X). Instead, we focus on the sign and the statistical significance of these coefficients, in particular, the "slopes.“

If a "slope" is positive, it means that an increase in the corresponding X variable increases the latent propensity to choose the 1 alternative, thereby also increasing the predicted probability of choosing the 1 alternative. (Analogously for negative slope coefficients.)

If the fitted probability is greater than 0.5, we predict the individual will choose 1. If it is less than 0.5, we predict they will choose 0. The asymptotic t-ratios allow us to test zero hypotheses about these "slopes." If the t-ratios exceed (roughly) 2 in absolute value, we can reject the hypothesis that the relevant X variable has no effect on fitted choice probabilities.

Missing Observations Collect data on Sleep and Age All data on Sleep but 20% of Age is missing How do you use all the data to show the effect of Age on Sleep?

Non-Experimental Data Non-experimental data can sometimes make it very difficult to draw policy implications from regression analysis

GUN CONTROL Suppose your sample consists of households that have been victimized by robbery. The dependent variable takes a value of 1 if a household member is shot during the robbery and 0 otherwise. One of your explanatory variables is a dummy variable equal to 1 if there is a handgun present in the house, 0 otherwise. When a handgun is present in a household, an occupant of that house is much more likely to be shot in the process of a robbery than when no handgun is present. Therefore, to minimize injury and loss of life from robbery incidents, private ownership of handguns should be banned. Evaluate this policy proposal and the "evidence" upon which it is premised

Briefly describe the nature of the true "experiment" that would allow an unambiguous determination of the effect of handgun presence on robbery shootings via a regression like this.

LEGALIZATION OF MARIJUANA: Suppose you have a random sample of at-risk 18-year- olds. The dependent variable is the number of times each teenager has used heroin. Among the explanatory variables is a dummy variable that takes a value of 1 if the subject experimented with marijuana prior to age 13, and 0 otherwise. You find that the coefficient on this dummy variable is positive and strongly statistically significant. Therefore, we should not legalize marijuana use (which would make it much more accessible to pre- teens) since this will lead to widespread use of heroin. Evaluate this policy proposal and the "evidence" upon which it is premised

Briefly describe the nature of the true "experiment" that would allow an unambiguous determination of the effect of pre-teen marijuana use on subsequent heroin use via a regression like this