Jean Yoon July 11, 2012 Research Design. Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences.

Slides:



Advertisements
Similar presentations
Econometric analysis informing policies UNICEF workshop, 13 May 2008 Christian Stoff Statistics Division, UNESCAP,
Advertisements

PANEL DATA 1. Dummy Variable Regression 2. LSDV Estimator
Economics 20 - Prof. Anderson1 Panel Data Methods y it = x it k x itk + u it.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
REGRESSION, IV, MATCHING Treatment effect Boualem RABTA Center for World Food Studies (SOW-VU) Vrije Universiteit - Amsterdam.
Advantages and limitations of non- and quasi-experimental methods Module 2.2.
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
Random Assignment Experiments
Presented by Malte Lierl (Yale University).  How do we measure program impact when random assignment is not possible ?  e.g. universal take-up  non-excludable.
Differences-in-Differences
The World Bank Human Development Network Spanish Impact Evaluation Fund.
PHSSR IG CyberSeminar Introductory Remarks Bryan Dowd Division of Health Policy and Management School of Public Health University of Minnesota.
Impact Evaluation Methods. Randomized Trials Regression Discontinuity Matching Difference in Differences.
Pooled Cross Sections and Panel Data II
Impact Evaluation: The case of Bogotá’s concession schools Felipe Barrera-Osorio World Bank 1 October 2010.
1Prof. Dr. Rainer Stachuletz Panel Data Methods y it =  0 +  1 x it  k x itk + u it.
Single and Multiple Spell Discrete Time Hazards Models with Parametric and Non-Parametric Corrections for Unobserved Heterogeneity David K. Guilkey.
1 Lecture 20: Non-experimental studies of interventions Describe the levels of evaluation (structure, process, outcome) and give examples of measures of.
Tax Subsidies for Out-of-Pocket Healthcare Costs Jessica Vistnes Agency for Healthcare Research and Quality William Jack Georgetown University Arik Levinson.
Econometrics with Observational Data: Research Design Todd Wagner.
Non Experimental Design in Education Ummul Ruthbah.
Qualitative and Limited Dependent Variable Models ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Matching Methods. Matching: Overview  The ideal comparison group is selected such that matches the treatment group using either a comprehensive baseline.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
CAUSAL INFERENCE Shwetlena Sabarwal Africa Program for Education Impact Evaluation Accra, Ghana, May 2010.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Jean Yoon December 15, 2010 Research Design. Outline Causality and study design Quasi-experimental methods for observational studies –Covariate matching.
Africa Impact Evaluation Program on AIDS (AIM-AIDS) Cape Town, South Africa March 8 – 13, Causal Inference Nandini Krishnan Africa Impact Evaluation.
Public Policy Analysis ECON 3386 Anant Nyshadham.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
Christine Pal Chee October 9, 2013 Research Design.
Causal Inference: experimental and quasi-experimental methods Draft ©G. Mason (2005)
Applying impact evaluation tools A hypothetical fertilizer project.
Non-experimental methods Markus Goldstein The World Bank DECRG & AFTPM.
Chapter 10 Finding Relationships Among Variables: Non-Experimental Research.
Chapter 6 Introduction to Multiple Regression. 2 Outline 1. Omitted variable bias 2. Causality and regression analysis 3. Multiple regression and OLS.
Lorraine Dearden Director of ADMIN Node Institute of Education
Applying Causal Inference Methods to Improve Identification of Health and Healthcare Disparities, and the Underlying Mediators and Moderators of Disparities.
Randomized Assignment Difference-in-Differences
Bilal Siddiqi Istanbul, May 12, 2015 Measuring Impact: Non-Experimental Methods.
1 Empirical methods: endogeneity, instrumental variables and panel data Advanced Corporate Finance Semester
Lecture 1 Introduction to econometrics
Summary: connecting the question to the analysis(es) Jay S. Kaufman, PhD McGill University, Montreal QC 26 February :40 PM – 4:20 PM National Academy.
MATCHING Eva Hromádková, Applied Econometrics JEM007, IES Lecture 4.
By Randall Munroe, xkcd.com Econometrics: The Search for Causal Relationships.
Impact Evaluation Methods Randomization and Causal Inference Slides by Paul J. Gertler & Sebastian Martinez.
The Evaluation Problem Alexander Spermann, University of Freiburg 1 The Fundamental Evaluation Problem and its Solution SS 2009.
Alexander Spermann University of Freiburg, SS 2008 Matching and DiD 1 Overview of non- experimental approaches: Matching and Difference in Difference Estimators.
Experimental Evaluations Methods of Economic Investigation Lecture 4.
Copyright © 2015 Inter-American Development Bank. This work is licensed under a Creative Commons IGO 3.0 Attribution-Non Commercial-No Derivatives (CC-IGO.
Module 5 Inference with Panel Data
Measuring Results and Impact Evaluation: From Promises into Evidence
12 Inferential Analysis.
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Matching Methods & Propensity Scores
Methods of Economic Investigation Lecture 12
Economics 20 - Prof. Anderson
1 Causal Inference Counterfactuals False Counterfactuals
Matching Methods & Propensity Scores
LIMITED DEPENDENT VARIABLE REGRESSION MODELS
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
12 Inferential Analysis.
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Explanation of slide: Logos, to show while the audience arrive.
Applying Impact Evaluation Tools: Hypothetical Fertilizer Project
Positive analysis in public finance
Ch. 13. Pooled Cross Sections Across Time: Simple Panel Data.
Presentation transcript:

Jean Yoon July 11, 2012 Research Design

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Causality Want to be able to understand impact of implementing new program or intervention. Ideally, would estimate causal effect of treatment on outcomes by comparing outcomes under counterfactual –Treatment effect=Y i (1)-Y i (0) –Observe outcome Y when patient gets treatment, t=1 and when same patient does not get treatment, t=0 –Compare difference in outcomes to get impact of treatment –In reality we don’t observe same patients with and without treatment

Randomized Experiment Randomize who gets treatment T RTO RO Compare outcome between treated and untreated groups to get impact of treatment Because treatment was randomized, there are no systematic differences between treated and untreated groups. Differences in outcomes can be attributed to causal effect of treatment

Causality and Observational Studies Most health services research is observational and cross sectional –Causality difficult to show because of confounding also referred to as selection and endogeneity  Omitted variables bias  Selection  Reverse causality  Measurement error

Observational Study Example Observe some patients with diabetes in primary care clinic participate in phone-based, disease management program, others don’t. –Compare A1c, cholesterol, other outcomes between groups of patients at end of program –If patients who participated in the program had better outcomes than those who didn’t, can we conclude the program caused the better outcomes?

Q&A Please respond using Q&A panel: –What other factors could have led the program participants to have better outcomes than the non-participants?

Bias of Treatment Effect  Characteristics not be balanced between groups  Enrolled had better outcomes to start with  Patients selected into treatment  Enrolled would have improved over time b/c more motivated to improve  Changes over time occurred that were unrelated to intervention  Enrolled also engaged in other activities (not part of program), e.g. reading diabetes info online

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Quasi-experimental Methods Observational studies do not have randomized treatment, so use methods to make like experimental study. –Identify similar control group –Try to eliminate any systematic differences between treatment and control groups  Compare (change in) outcomes between treatment and control groups

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Sample selection Bias arises when outcome is not observable for some people, sample selection Form of omitted variables bias Example: attrition from study collecting outcomes data from patients –Health outcomes only observed in respondents –Not possible to make inferences about determinants of health outcomes in study population as whole. –If reasons why patients don’t respond are correlated with unobservable factors that influence outcomes, then identification problem.

Sample selection One solution to sample bias is a selection model called Heckman probit or heckit or generalized tobit First stage uses probit model for dichotomous outcome (selection equation) for whole sample From parameters in first stage, calculate inverse mills ratio or selection hazard for each patient Add inverse mills ratio as variable into second stage Second stage uses OLS model for linear outcome (outcome equation) for respondents in sample Assumption is that error terms of two equations are jointly normally distributed and correlated by ρ

Sample selection 1) D=Zγ+ε D=1 participation, D=0 no participation 2) w*=Xβ+μ Only have outcome w* when D=1 Estimate 1) and get φ(Zγ)/Φ(Zγ)= λ inverse mills ratio Estimate 2) adding λ as a parameter E[w|X, D=1]= Xβ + ρδ μ λ(Zγ) Coefficient of λ is ρ, if =0 then no sample selection Want instruments (included in Z) that are not related to outcome w

Sample selection If outcome is dichotomous, can use bivariate probit model. If outcomes data are available for all patients, then use treatment effects model. New, semiparametric models do not make joint normality assumptions.

Sample selection Strengths –Generalizes to population of interest –Addresses selection based on unobservables Weaknesses –Assumes joint normality of errors in two equations. –Can still be run if no instrument (unlike 2SLS) but relying on distributional assumptions of non- linearity in inverse mills ratio.

Sample selection Stata code Heckman y x1 x2 x3, select (x1 x2 z1 z2) twostep

Sample Selection Example Zhoua A, Gaob J, Xueb Q, Yangb X, Yan J. Effects of rural mutual health care on outpatient service utilization in Chinese village medical institutions: evidence from panel data. Health Econ. 18: S129– S136 (2009) Compare effect of outpatient insurance program and drug policy to restrict drug prescribing on outpatient medical expenditures. Only some residents had outpatient visits.

Sample Selection Example Zhou et al. results

Sample Selection Example Zhou et al. results

Poll 1

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Differences-in-Differences Can exploit natural experiment with D-D Need longitudinal data or observe outcome at different time points for treatment and control groups Subtract out differences between treatment and control groups and differences over time

Differences-in-Differences Average treatment effect=(B-A) – (D-C)

Differences-in-Differences  Program P: 0=no, 1=yes  Time T: 0=pre, 1=post Y= β 0 + β 1 T + β 2 P + β 3 P*T +ε So β 3 is the differences-in-differences estimate

Differences-in-Differences Strengths –Difference out time trend –Addresses omitted variables bias if unmeasured time invariant factors Weaknesses –If unobserved factors that change over time, can have biased estimates

Differences-in-Differences  Unobserved factors often cause omitted variables bias  Panel analysis with time invariant characteristics δ i for individual (observed and unobserved) Y it = β 0 + β 1 T t + β 2 P it +δ i +ε it  Difference model Y i1 -Y i0 = β 1 + (P i1 -P i0 )*β 2 + ε i1 –ε i0  β 1 is time trend  β 2 is treatment effect  Time invariant δ i drops out of model

Differences-in-Differences Fixed effects estimate of “within estimator” same as first differencing with 2 time periods Cannot estimate effect of time invariant factors in FE model Stata command xtreg will run FE

D-D Example Chernew ME et al. Impact Of Decreasing Copayments On Medication Adherence Within A Disease Management Environment. Health Affairs; Jan/Feb 2008; 27, 1; Two health plans implemented disease management programs, but only one reduced drug copayments Drug adherence for two groups of patients compared pre-post implementation of disease management and reduction in drug copays

D-D Example Chernew et al.

Poll 2

Outline Causality and study design Quasi-experimental methods for observational studies –Sample selection –Differences-in-differences –Regression discontinuity

Regression Discontinuity Can do when treatment is not randomly assigned but based on a continuous, measured factor Z –Z called forcing variable Discontinuity at some cutoff value of Z Individuals cannot manipulate assignment of Z Only jump in outcome due to discontinuity of treatment Treatment effect =the expected outcome for units just above the cutoff minus the expected outcome for units just below the cutoff (otherwise identical)

Regression Discontinuity Strengths –Z can have direct impact on outcome (unlike instrumental variables) Weaknesses –Need to test functional form for effect of treatment (e.g. linear, interaction, quadratic terms) or can get biased treatment effects if model is misspecified.

RD Example Bauhoff, S., Hotchkiss, D. R. and Smith, O. (2011), The impact of medical insurance for the poor in Georgia: a regression discontinuity approach. Health Econ., 20: 1362–1378. Effect of medical insurance program for poor in republic of Georgia on utilization Eligibility for program limited to residents below means test score (SSA) Compare outcomes for eligible residents versus low income residents who are not eligible

RD Example Bauhoff et al

RD Example Bauhoff et al Y=β 0 +β 1 MIP+β 2 f(score-cutoff)+β 3 MIP*f(score-cutoff)+β 4 X+ε β 1 =treatment effect, discontinuous change at cutoff β 2 =effect of means test on outcomes for non-beneficiaries β 3 =effect of means test on outcomes for beneficiaries

RD Example Bauhoff et al

Poll 3

Poll 4

Review Quasi-experimental methods can help address common sources of bias of treatment effects in observational studies. Quasi-experimental methods attempt to reduce any systematic differences between treatment and control groups. Quasi-experimental methods provide stronger study designs in order to make inferences about causality.

References Campbell, D. T., and Stanley, J. C. Experimental and Quasi- experimental Designs for Research. Chicago: Rand McNally, Heckman, J. (1979). "Sample selection bias as a specification error". Econometrica 47 (1): 153–61. Wooldridge, J. M.: Econometric Analysis of Cross Section and Panel Data. MIT Press, Cambridge, Mass., Trochim, William M. The Research Methods Knowledge Base, 2nd Edition. Internet WWW page, at URL: current as of 12/07/10)./

More References Zhoua A, Gaob J, Xueb Q, Yangb X, Yan J. Effects of rural mutual health care on outpatient service utilization in Chinese village medical institutions: evidence from panel data. Health Econ. 18: S129–S136 (2009) Chernew ME et al. Impact Of Decreasing Copayments On Medication Adherence Within A Disease Management Environment. Health Affairs; Jan/Feb 2008; 27, 1; Bauhoff, S., Hotchkiss, D. R. and Smith, O. The impact of medical insurance for the poor in Georgia: a regression discontinuity approach. Health Econ. 2011; 20: 1362–1378.

Next Lectures July 25, 2012 Propensity Scores Todd Wagner August 8, 2012 Instrumental Variables Models Patsi Sinnott