Drop-in Sessions! When: Hillary Term - Week 1 Where: Q-Step Lab (TBC) Sign up with Alice Evans.

Slides:



Advertisements
Similar presentations
Dummy Dependent variable Models
Advertisements

© Department of Statistics 2012 STATS 330 Lecture 32: Slide 1 Stats 330: Lecture 32.
Brief introduction on Logistic Regression
Logistic Regression Psy 524 Ainsworth.
Logistic Regression.
Nguyen Ngoc Anh Nguyen Ha Trang
Multiple Linear Regression Model
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
QUALITATIVE AND LIMITED DEPENDENT VARIABLE MODELS.
Log-linear and logistic models Generalised linear model ANOVA revisited Log-linear model: Poisson distribution logistic model: Binomial distribution Deviances.
Statistical Analysis SC504/HS927 Spring Term 2008 Session 7: Week 23: 7 th March 2008 Complex independent variables and regression diagnostics.
Chapter 11 Multiple Regression.
(Correlation and) (Multiple) Regression Friday 5 th March (and Logistic Regression too!)
An Introduction to Logistic Regression
Business Statistics - QBM117 Statistical inference for regression.
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
1 Regression Models with Binary Response Regression: “Regression is a process in which we estimate one variable on the basis of one or more other variables.”
9 - 1 Intrinsically Linear Regression Chapter Introduction In Chapter 7 we discussed some deviations from the assumptions of the regression model.
Introduction to Linear Regression and Correlation Analysis
Logistic Regression- Dichotomous Dependent Variables March 21 & 23, 2011.
ALISON BOWLING THE GENERAL LINEAR MODEL. ALTERNATIVE EXPRESSION OF THE MODEL.
Logistic Regression STA2101/442 F 2014 See last slide for copyright information.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 7 Logistic Regression I.
When and why to use Logistic Regression?  The response variable has to be binary or ordinal.  Predictors can be continuous, discrete, or combinations.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Regression. Types of Linear Regression Model Ordinary Least Square Model (OLS) –Minimize the residuals about the regression linear –Most commonly used.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
Logistic Regression. Linear Regression Purchases vs. Income.
Multiple Logistic Regression STAT E-150 Statistical Methods.
1 Topic 4 : Ordered Logit Analysis. 2 Often we deal with data where the responses are ordered – e.g. : (i) Eyesight tests – bad; average; good (ii) Voting.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Regression Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
4. Binary dependent variable Sometimes it is not possible to quantify the y’s Ex. To work or not? To vote one or other party, etc. Some difficulties: 1.Heteroskedasticity.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
Logistic Regression and Odds Ratios Psych DeShon.
Nonparametric Statistics
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Multiple Regression Analysis Bernhard Kittel Center for Social Science Methodology University of Oldenburg.
Logistic Regression: Regression with a Binary Dependent Variable.
Nonparametric Statistics
BINARY LOGISTIC REGRESSION
Chapter 4 Basic Estimation Techniques
Statistical Modelling
Logistic Regression When and why do we use logistic regression?
A priori violations In the following cases, your data violates the normality and homoskedasticity assumption on a priori grounds: (1) count data  Poisson.
Regression Analysis AGEC 784.
Logistic Regression APKC – STATS AFAC (2016).
Advanced Quantitative Techniques
Notes on Logistic Regression
Advanced Quantitative Techniques
THE LOGIT AND PROBIT MODELS
BIVARIATE REGRESSION AND CORRELATION
Introduction to logistic regression a.k.a. Varbrul
Regression Analysis Week 4.
THE LOGIT AND PROBIT MODELS
Model Diagnostics and OLS Assumptions
Nonparametric Statistics
Correlation and Regression-III
Introduction to Logistic Regression
Simple Linear Regression
What’s the plan? First, we are going to look at the correlation between two variables: studying for calculus and the final percentage grade a student gets.
Chapter 6 Logistic Regression: Regression with a Binary Dependent Variable Copyright © 2010 Pearson Education, Inc., publishing as Prentice-Hall.
Logistic Regression.
Generalized Additive Model
Regression Part II.
Nazmus Saquib, PhD Head of Research Sulaiman AlRajhi Colleges
Presentation transcript:

Drop-in Sessions! When: Hillary Term - Week 1 Where: Q-Step Lab (TBC) Sign up with Alice Evans

Political Analysis 2, Week 8, Lab 7 Logistic Regression Political Analysis 2, Week 8, Lab 7

Recap Linear regression models are useful for approximating many relationships between variables. Ex: 1. The effect of foreign aid on GDP p.c. (Easterly 2003) Ex: 2. The effect of democracy on trade liberalization (Milner & Kubota 2005) But some relationships violate the assumption of linearity, i.e., the marginal effect of X (our IV) on Y (our DV) is not constant. Exponential Logarithmic Cubic Quadratic

Transforming variables Recall from last week… Non-linear relationships between X and Y can be modelled by transforming X . Ex: Y= B0 + B1*X + B2*X2 + B3*logeZ

Changing the model Another option at our disposal is to change the statistical model to better estimate the relationship between Xs and Y. The model we choose will depend on the nature of our variables Focus for today: the dependent variable is binary (Y=0 or Y=1) Variable Y=1 Y=0 Legislation bill passes bill fails Conflict war peace IO participation member not a member Climate change support signatory to Paris non-signatory to Paris Legislative influence party in majority party not in majority Civil war onset no onset

Changing the model: LPM Some options: OLS Linear Probability Model Assumptions of ordinary least squares apply e.g., DV is a linear function of the IVs; normal distribution of errors; etc. But DV is modelled differently! Y = Pr(event occurring) Pr 𝑌 = 1 = 𝑏 0 + 𝑏 1 𝑋 1 + 𝑏 2 𝑋 2 +𝑒 Beta interpretation is straightforward!: change in Pr(Y = 1) following a one-unit change in X, holding all other independent variables constant

Changing the model: Logit Some options: 2. Logistic regression While OLS models the relationship btw IVs and the DV as a straight line (with predicted Ys btw -∞ and +∞)… logistic regression models the relationship as a smooth curve (with predicted Ys btw 0 and 1). Logit models allow us to assume a certain king of nonlinear relationship btw the IVs and the Pr(Y=1)

Linear to logistic Some drawbacks of LPM Potential for out-of-bound predictions (recall Olympic 100 m sprint example) Non-normality of errors (binomial instead), which can invalidate t-tests Heteroscedasticity (variance of residuals is not constant

Probability and odds Logistic regression allows us to determine the likelihood of an outcome as measured by the probability or odds it occurs Probability (Pr) Range is 0 to 1 Pr(Y=1) + Pr(Y=0)) = 1 Ex: If the probability of Art. 50 being invoked in 2017 is 20% (.2) then the probability of Art. 50 not being invoked is 80% (.8) Odds Range is 0 to +∞ Odds is equal to Pr(Y=1)/Pr(Y=0) or Pr(Y=1)/1-Pr(Y=1) Ex: Odds of invoking Art. 50 in 2017 are .2/.8 = .25 = the odds of Art. 50 being invoked are 1:4.

Probability and odds Range: 0 to +∞ Range: 0 to 1 The relationship between odds and probability is monotonic. As odds , probability , and vice versa

Logistic regression The left side of our logistic regression equation represents the “logit.” The logit is the transformation of Y from odds to the log of odds. (This log transformation is also monotonic.)

Logistic regression cont. How to interpret logit coefficients? Log of odds ratio Means a one-unit increase in β (warl) reduces the log of odds ratio for civil war onset by .95437.

Interpreting logit coefficients To interpret logit coefficients, we have two options: 1. Log of odds ratio  (conditional) odds ratio Take the coefficient’s exponential and subtract from 1 Gives us the percentage change in odds following a one-unit increase in X 2. Generate predicted probabilities (focus of today) Specify range of values for an IV of interest Specify constant values for non-predicted IVs Store in a dataframe  predict() Probabilities from 0 to 1

Interpreting logit coefficients cont. Critically… …whether coefficients are interpreted using odds or probabilities… ...the marginal effect of any independent variable depends on the value of that independent variable as well as the values of other covariates in the model.

Evaluation Please take a few minutes at the end of class to evaluate this term’s lab! https://www.surveymonkey.co.uk/r/QS2_MT17