FINAL MEETING – OTHER METHODS Development Workshop.

Slides:



Advertisements
Similar presentations
Introduction to Propensity Score Matching
Advertisements

REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Analyzing Regression-Discontinuity Designs with Multiple Assignment Variables: A Comparative Study of Four Estimation Methods Vivian C. Wong Northwestern.
Regression Discontinuity (Durham, 8 April 2013) Hans Luyten University of Twente, Faculty of Behavioural Sciences.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Copyright (c) Bani Mallick1 Lecture 2 Stat 651. Copyright (c) Bani Mallick2 Topics in Lecture #2 Population and sample parameters More on populations.
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
Business and Economics 7th Edition
© Institute for Fiscal Studies The role of evaluation in social research: current perspectives and new developments Lorraine Dearden, Institute of Education.
Mitchell Hoffman UC Berkeley. Statistics: Making inferences about populations (infinitely large) using finitely large data. Crucial for Addressing Causal.
Impact Evaluation Session VII Sampling and Power Jishnu Das November 2006.
SHOWTIME! STATISTICAL TOOLS IN EVALUATION DESCRIPTIVE VALUES MEASURES OF VARIABILITY.
Random Variables and Probability Distributions
Basic Analysis of Variance and the General Linear Model Psy 420 Andrew Ainsworth.
Bootstrapping applied to t-tests
The mathematics skills of school children: How does England compare to the high performing East Asian jurisdictions?
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
Describing distributions with numbers
Non Experimental Design in Education Ummul Ruthbah.
POSC 202A: Lecture 2 Homework #1: 1.2, 1.44, 1.54, 1.62,1.74, 3.2, 3.6, 3.52, 3.54, 3.60, 3.67, 3.70 Today: Research Designs, Mean, Variance.
Using the Margins Command to Estimate and Interpret Adjusted Predictions and Marginal Effects Richard Williams
Bootstrapping (And other statistical trickery). Reminder Of What We Do In Statistics Null Hypothesis Statistical Test Logic – Assume that the “no effect”
PARAMETRIC STATISTICAL INFERENCE
1 Psych 5500/6500 Standard Deviations, Standard Scores, and Areas Under the Normal Curve Fall, 2008.
Linear Regression Model In regression, x = independent (predictor) variable y= dependent (response) variable regression line (prediction line) ŷ = a +
Estimating Causal Effects from Large Data Sets Using Propensity Scores Hal V. Barron, MD TICR 5/06.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Generalizing Observational Study Results Applying Propensity Score Methods to Complex Surveys Megan Schuler Eva DuGoff Elizabeth Stuart National Conference.
Applying impact evaluation tools A hypothetical fertilizer project.
New research methods for the evaluation of policy changes Sanjay Basu, MD, PhD O LD P ROBLEMS, N EW S OLUTIONS.
Comparing Two Means Chapter 9. Experiments Simple experiments – One IV that’s categorical (two levels!) – One DV that’s interval/ratio/continuous – For.
Africa Program for Education Impact Evaluation Dakar, Senegal December 15-19, 2008 Experimental Methods Muna Meky Economist Africa Impact Evaluation Initiative.
Lorraine Dearden Director of ADMIN Node Institute of Education
Using Propensity Score Matching in Observational Services Research Neal Wallace, Ph.D. Portland State University February
AADAPT Workshop Latin America Brasilia, November 16-20, 2009 Laura Chioda.
Robust Regression. Regression Methods  We are going to look at three approaches to robust regression:  Regression with robust standard errors  Regression.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
Causal inferences This week we have been discussing ways to make inferences about the causal relationships between variables. One of the strongest ways.
Non-parametric Approaches The Bootstrap. Non-parametric? Non-parametric or distribution-free tests have more lax and/or different assumptions Properties:
Modern Approaches The Bootstrap with Inferential Example.
Summary: connecting the question to the analysis(es) Jay S. Kaufman, PhD McGill University, Montreal QC 26 February :40 PM – 4:20 PM National Academy.
MATCHING Eva Hromádková, Applied Econometrics JEM007, IES Lecture 4.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Empirical Studies of Marriage and Divorce. Korenman and Neumark, 1991 Does Marriage Really Make Men More Productive? How do we explain the male marriage.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
Review Design of experiments, histograms, average and standard deviation, normal approximation, measurement error, and probability.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
ENDOGENEITY - SIMULTANEITY Development Workshop. What is endogeneity and why we do not like it? [REPETITION] Three causes: – X influences Y, but Y reinforces.
Impact Evaluation Methods Regression Discontinuity Design and Difference in Differences Slides by Paul J. Gertler & Sebastian Martinez.
Estimating standard error using bootstrap
Lezione di approfondimento su RDD (in inglese)
An introduction to Impact Evaluation
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Impact Evaluation Methods: Difference in difference & Matching
Basic Practice of Statistics - 3rd Edition Two-Sample Problems
Evaluating Impacts: An Overview of Quantitative Methods
Analysing RWE for HTA: Challenges, methods and critique
Random Variables and Probability Distributions
Density Curves Normal Distribution Area under the curve
Using the Rule Normal Quantile Plots
Presentation transcript:

FINAL MEETING – OTHER METHODS Development Workshop

General conclusions on causal analyses Magic tool of „ceteris paribus” – Regression is ceteris paribus by definition – But the data need not to be – they are just a subsample of general populations and many other things confound Causal effects, i.e. cause and effect – Propensity Score Matching – Regression Discontinuity – Fixed Effects – Instrumental Variables 2

If we cannot experiment..… 3 Cross-sectional data Panel data „Regression Discontinuity Design“ „Propensity Score Matching“ IV Before After Estimators Difference in Difference Estimators (DiD) „Propensity Score Matching“ + DiD

Problems with causal inference 4 Confounding Influence (environment) Treatment Effect Observables Unobservables

Instrumental Variables solution… Confounding Influence Treatment Outcome Instrumental Variable(s) Observed Factor Unobserved Factor

Fixed Effects Solution… (DiD does pretty much the same) Confounding Influence Treatment Outcome Fixed Influences Observed Factor Unobserved Factor

Propensity Score Matching Confounding Influence Treatment Outcome Treatment Observed Factor Unobserved Factor

Regression Discontinuity Design Confounding Influence Treatment Effect Group that is key for this policy Observables Unobservables 8

A motivating story Today women in Poland have on average 1,7 kid About 50 years ago, women had 2,8 kids Todays women are 6 times more educated than 50 years ago – will a drop from 2.8 to 1.7 be an effect of this educational change? Natural experiment: in 1960 schooling obligation was extended by one year (11 to 12 years). – THE SAME women born just before 1953 went to primary and secondary schools a year shorter than born after 1953 – THE SAME = ? RD allows to compare fertility (with individual characteristics) for women born around

Regression Discontinuity Design Idea – Focus your analyses on a group for which treament was random (or rather: independent) How to do it? – Example: weaker students have lower grades, but are also frequently „delayed” to repeat courses/years; if we give them extra classes, better students will outperform them anyway, so how to test if extra classes help? – RDD will compare the performance of students just above and just below „threshold”, so quite similar ones – RDD will only work if people cannot „prevent” or „encourage” treatment by relocating themselves around „threshold” 10

Regression Discontinuity Design Advantages: – Really marginal effect – Causal, if RDD well applied Disadvantages: – Sample size largely limited – Only „local” character of estimations (marginal≠average) Problems: – How do we know how far away from threshold can we go (bandwidth)? – How do we know if design is ok.? 11

Regression Discontinuity Design Zastosowanie – Trade off between narrow “bandwidth” (for independence assumption) and wide “bandwidth” to increase sample size – One can try to find it empirically ( “fuzzy” RD design) – Y is the effect, p is treatment probability. + is effect of probability just above „cut-off” - is effect of probability just below „cut-off” 12

Regression Discontinuity Design 13

Regression Discontinuity Design 14

Regression Discontinuity Design 15

16 How to do this in STATA? First – download package: net instal rd Second – define your model – rd $out, treatment, $in [if] [in] [weight] [, options] Third – there are some options – mbw(numlist) multiplication of „bandwidth” in percent (default: " " which means we always do 50%, 100% and 200%) – z0(real) sets cutoff Z0 (treatment) – ddens asks for extra estimation of discontinuities in Z density – graph – draws graphs we’ve seen automatically

Sample results in STATA - data

Output from STATA 18

Output from STATA - graph

Output from STATA –„fuzzy” version 20 gen byte ranwin=cond(uniform()<.1,1-win,win) rd lne ranwin d, mbw(25(25)300) bdep ox

Quintile regressions One last thing

A motivating story

Some basics „doubts” of an empirical economist… Compare similar to similar Keep statistical properties Understand bezond „average x” Understand (and be independent of) „outliers”

Robust estimators First flavour of robust – regression with robust option – Helps if problem is not systematic – Does not help if problem is the nature of the process (e.g. heterogeneity) Second flavour of robust – nonparametric estimators – Complex from mathematical point of view – Takes longer to compute – But veeeery elastic => Koenker (and his followers)

How to do this in STATA? Estimate at median – qreg y $in Estimate at any other percentile – qreg y $in, quantile(q) where q is your percentile Estimate differences between different percentiles – iqreg y $in, quantile(.25.75) reps(100) + additionally may bootstrap

Output from STATA

Summarising all this crap Confounding Influence (environment) Treatment Effect Observables Unobservables

Problems Sample – size – heterogeneity Methods – None is perfect – Question important – Nonparametric (kernel in PSM or QR) are robust, robust is not a synonim for miraculous