Evaluation Design. By the end of the session, participants will be a able to: 1.Name the criteria for inferring causality 2.Understand internal & external.

Slides:



Advertisements
Similar presentations
Andrea M. Landis, PhD, RN UW LEAH
Advertisements

Agency for Healthcare Research and Quality (AHRQ)
Experimental and Ex Post Facto Designs
Causality Causality Hill’s Criteria Cross sectional studies.
GROUP-LEVEL DESIGNS Chapter 9.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Experimental Research Designs
Program Evaluation Regional Workshop on the Monitoring and Evaluation of HIV/AIDS Programs February 14 – 24, 2011 New Delhi, India.
Reading the Dental Literature
Correlation AND EXPERIMENTAL DESIGN
Studying Behavior. Midterm Review Session The TAs will conduct the review session on Wednesday, October 15 th. If you have questions, your TA and.
Monitoring and Evaluation: Evaluation Designs. Objectives of the Session By the end of this session, participants will be able to: Understand the purpose,
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
9 Quantitative Research Designs.
Chapter 9 Experimental Research Gay, Mills, and Airasian
McGraw-Hill © 2006 The McGraw-Hill Companies, Inc. All rights reserved. Experimental Research Chapter Thirteen.
Experimental Research
Chapter 8 Experimental Research
Experimental Design The Gold Standard?.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Experimental and Quasi-Experimental Designs
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Selecting a Research Design. Research Design Refers to the outline, plan, or strategy specifying the procedure to be used in answering research questions.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Primary Data Collection: Experimentation CHAPTER eight.
Quantitative Research Designs
Day 6: Non-Experimental & Experimental Design
Epidemiology The Basics Only… Adapted with permission from a class presentation developed by Dr. Charles Lynch – University of Iowa, Iowa City.
Learning Objectives Copyright © 2004 John Wiley & Sons,Inc Primary Data Collection: Experimentation CHAPTER Seven.
Introduction to research Research designs Dr Naiema Gaber.
ECON ECON Health Economic Policy Lab Kem P. Krueger, Pharm.D., Ph.D. Anne Alexander, M.S., Ph.D. University of Wyoming.
Copyright ©2008 by Pearson Education, Inc. Pearson Prentice Hall Upper Saddle River, NJ Foundations of Nursing Research, 5e By Rose Marie Nieswiadomy.
1 Experimental Research Cause + Effect Manipulation Control.
STUDYING BEHAVIOR © 2009 The McGraw-Hill Companies, Inc.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
After giving this lecture the student should be able to do the following: After giving this lecture the student should be able to do the following: List.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
Chapter 10 Finding Relationships Among Variables: Non-Experimental Research.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
Experimental & Quasi-Experimental Designs Dr. Guerette.
SOCW 671: #6 Research Designs Review for 1 st Quiz.
The Experiment Chapter 7. Doing Experiments In Everyday Life Experiments in psychology use the same logic that guides experiments in biology or engineering.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Handbook for Health Care Research, Second Edition Chapter 7 © 2010 Jones and Bartlett Publishers, LLC CHAPTER 7 Designing the Experiment.
Understanding Quantitative Research Design
Experimental and Ex Post Facto Designs
CJ490: Research Methods in Criminal Justice UNIT #4 SEMINAR Professor Jeffrey Hauck.
Introduction to Monitoring and Evaluation. Learning Objectives By the end of the session, participants will be able to: Define program components Define.
Session 7: Planning for Evaluation. Session Overview Key definitions:  monitoring  evaluation Process monitoring and process evaluation Outcome monitoring.
© 2011 Pearson Education, Inc. All rights reserved. This multimedia product and its contents are protected under copyright law. The following are prohibited.
Criminal Justice and Criminology Research Methods, Second Edition Kraska / Neuman © 2012 by Pearson Higher Education, Inc Upper Saddle River, New Jersey.
MEASURE Evaluation Data Quality Assurance Workshop Session 3 Introduction to Routine Data Quality Assessment.
Experimental Design Ragu, Nickola, Marina, & Shannon.
William M. Trochim James P. Donnelly Kanika Arora 8 Introduction to Design.
Research Designs for Explanation Experimental, Quasi-experimental, Non-experimental, Observational.
Chapter 11: Quasi-Experimental and Single Case Experimental Designs
EXPERIMENTAL RESEARCH
Right-sized Evaluation
Experimental Research Designs
Making Causal Inferences and Ruling out Rival Explanations
Introduction to Design
Experiments and Quasi-Experiments
Quasi-Experimental Design
Experiments and Quasi-Experiments
Group Experimental Design
Measuring Data Quality
Primary Data Collection: Experimentation
Reminder for next week CUELT Conference.
Presentation transcript:

Evaluation Design

By the end of the session, participants will be a able to: 1.Name the criteria for inferring causality 2.Understand internal & external validity 3.Know the different types of designs for an evaluation 4.Identify the strengths and limitations of the different types of study designs 5.Develop an evaluation framework 6.Select a study design that fits the purpose of a given evaluation Learning Objectives

What are some different types of evaluation? Formative evaluation Process evaluation Impact evaluation

Introduction To show impact, researchers often want to make inferences about cause and effect We may, for instance, want to: –Identify the factors that explain why the use of ITNs is more prevalent among one group compared to another –Know why a particular intervention works

Logic of Causal Inference Under what condition may we infer that a change in dependent variable was really caused by the independent variable, and not by something else? What are some of the most plausible rival explanations, and how do we rule them out?

Overview In this section, we will examine appropriate and inappropriate criteria for inferring causality We will identify various evaluation designs: –Experimental –Quasi-experimental –Non-Experimental and the ways they attempt to show causality

C riteria for Inferring Causality: Temporal Relationship –The cause must precede the effect

Criteria for Inferring Causality: Plausibility –An association is plausible, and thus more likely to be causal, if consistent with other knowledge

Criteria for Inferring Causality: Strength of the Association –A strong association between possible cause and effect, as measured by the size of the relative risk, is more likely to be causal than a weak association

Criteria for Inferring Causality: Consistency Several studies giving the same result –Clearest when variety of study designs are used in different settings The likelihood that all studies are making the same mistake is minimized –Lack of consistency does not exclude a causal association Different exposure levels and other conditions may reduce the impact of the causal factor in certain studies

Criteria for Inferring Causality: Dose-response Relationship Occurs when changes in the level of a possible cause are associated with changes in the prevalence or incidence of the effect –The prevalence of hearing loss increases with noise level and exposure time

Criteria for Inferring Causality: Reversibility When the removal of a possible cause results in a reduced disease risk, the likelihood of the association being causal is strengthened –Cessation of cigarette smoking is associated with a reduction in the risk of lung cancer relative to that in people who continue to smoke However, when the cause leads to rapid irreversible changes whether or not there is continued exposure (HIV infection), then reversibility cannot be a condition for causality

Internal & External Validity

When considering cause and effect, two forms of validity become relevant: –Internal validity –External Validity

Internal Validity Internal validity: The confidence that the results of a study accurately depict whether one variable is or is not a cause of another –A study is internally valid if it meets our criteria: Cause precedes effect Empirical correlation between cause and effect No confounding variable

Threats to Internal Validity History Maturation or the passage of time Testing & Instrumentation Selection bias Loss to follow-up Diffusion or imitation of treatments

External Validity The extent to which the causal relationship depicted in a study can be generalized beyond the study conditions

Evaluating the Effect of an Intervention

Specific Evaluation Questions Service provisionAre the services available? Are they accessible? Is quality adequate? Service utilizationAre the services being used? Service coverageIs the target population being reached? Service effectiveness Impact Is there improvement in disease outcome or health-related behavior? Were the improvements due to the program?

Measuring the Effect of an Intervention Start of Intervention ends Time Change in Intervention outcome With intervention Without intervention Intervention impact

Deciding on Evaluation Design 1.What is your research question? 2.What is your target population? 3.What do you know about this population? 4.How do you intend to use the results? 5.What do you want to measure (indicators)? 6.What type of inference do you want to draw? 7.When do you need the results? 8.Do you have a sampling frame? What shape is it in? 9.How much are you willing to pay? 10.Where in the program life cycle are you now?

Types of Evaluation Design Design ExperimentalStrongest for demonstrating causality, most expensive Quasi- experimental Weaker for demonstrating causality, less expensive Non- experimental Weakest for demonstrating causality, least expensive

Experimental Design

The Basic Experimental Principle The intervention is the only difference between two groups This is achieved by random assignment

Pretest-posttest Experimental Design Experimental designs attempt to provide maximum control for threats to internal validity They do so by giving the researchers greater ability to manipulate and isolate the independent variable –(not always possible in practice)

Pretest-posttest Experimental Design Essential components: 1.Identify the study population and determine the appropriate sample size for experimental and control groups 2.Randomly assign individuals to experimental and control groups 3.Pre-test everyone with a standardized instrument 4.Introduce the independent variable (intervention) to the experimental group while withholding it from the control group 5.Post-test both groups with the same instrument and under the same conditions as the pretest 6.Compare the amount of change in dependent variable for both experimental and control groups

Pretest-posttest Control Group Design Intervention group Control group Program Pre- test Post- test Randomization Pre- test Time

Factors that May Distort Conclusions Dropout Instrumentation effects Testing effects If you think that taking the pretest might influence the treatment effects, or if it might bias the post-test responses, you might want to opt for the posttest- only control group design Contamination

Posttest-only Experimental Design

Post Test Only Experimental Design Initial group equivalence Differences between the experimental and control groups at posttest are assumed to reflect the causal impact of the independent variable

Post Test Only Experimental Design Intervention group Control group Program Post- test Randomization Time

Post-test Only, what to consider? Disadvantages Cannot assess whether the program is going to people for whom it was intended Cannot check comparability of groups Cannot know how much change actually occurred However, a pre-test post-test design always preferred Advantages Cheaper Useful when pre-test can interfere with program effects Randomization ensures equivalent experimental and control groups

Group Discussion In which situations might experimental design not be possible?

Possible Responses Randomization needed before program starts Ethics o Solution: Use alternative program rather than no program o Known efficacy of intervention Political factors Scale-up o Solution: Start out on small scale and use delayed program strategy

Quasi-Experimental Design

Quasi-experimental designs Can be used when random assignment is not possible Less internal validity than “true” experiments Still provide a moderate amount of support for causal inferences

Principles of the Quasi-experimental Design Intervention group Comparison group Program Pre- test Post- test Pre- test Time (Pre- and Post-test with comparison group but not randomized) Keep in mind selection effects: these occur when people selected for a comparison group differ from the experimental group

o Difference-in-Difference analysis: method involves comparing changes before and after the program for individuals in the program and control groups o Regression analysis: Attempts to address the problem of confounding by controlling for difference at baseline Quasi-experimental Statistical Methods

Summary of Quasi-Experimental Design Can demand more time and resources Require access to at least two similar groups Provides the assurance that outcomes are actually the results of the program Allows you to accurately assess how much of an effect the program has AdvantageDisadvantage

Non-Experimental Design

Pre-test Post-test Non-experimental design Intervention group Program Pre- test Post- test Time Pre-test Post-test (No comparison group )

Merits of Pre-test Post-test Non- experimental design Cannot account for non- program influences on outcomes Causal attribution not possible Cannot detect small but important changes If self-reporting is used rather than objective measures, posttest scores may be lower than pretest scores Relatively simple to implement Controls for participants' prior knowledge/attitudes/skills intentions AdvantageDisadvantage

Time Series Design Intervention group Program Pretest 2 Posttest 1 Time (No comparison group ) Pretest 1 Posttest 2

Merits of Time Series Design Problem of confounding Changes in instruments during the series of measurements Loss or change of cases Changes in group composition Enables detection of whether program effects are long-term or short-term Series of tests before intervention can eliminate need of control group Series of tests before program can be used to project results which would be expected Can be used if you have only one site to conduct your evaluation Advantage Disadvantage

Strengthening non-experimental designs Since there is no control group confounding can be a problem By constructing a plausibility argument and controlling for contextual and confounding factors, non-experimental designs can be strengthened

How to construct a Plausibility Argument Describe trends in Intervention coverage Intermediary outcomes Impact outcomes Contextual factors Link these trends –Temporal, spatial, age-pattern, “dose-response” associations

Plausibility Argument for Impact of Malaria Control Increase in effective intervention coverage Decreased morbidity Decreased malaria- associated mortality ITN Ownership ITN use IPTp Treatment Parasite prevalence Anaemia (<8g/dL) Fever All cause under-five mortality (5q0) Indicators Contextual factors Socioeconomic Education Fertility risk Housing condition Nutrition Climatic factor Rainfall Temperature Health intervention Health care utilization ANC, EPI, Vit A, PMTC

Merits of Plausibility Argument Need of several data points for intervention and outcomes Challenges in comparing data from different sources Non consistent data collection methods overtime Allow national scale program evaluation- Complex intervention Can accommodate the use of data from different sources AdvantageDisadvantage

Summary of Different Study Designs True experimentalQuasi-experimentalNon-experimental Partial coverage/ new programs Control group Strongest design Most expensive Partial coverage/ new programs Comparison group Weaker than experimental design Less expensive Full coverage programs -- Weakest design Least expensive More robust Less robust

Different designs vary in their capacity to produce information that allows for the linking of program outcomes to program activities The more confident you want to be about making these connections, the more rigorous the design and costly the evaluation Summary of Different Study Designs

Impact Evaluation Case Study 1: Reducing Malaria Transmission through IRS

Impact Evaluation Case Study 2: Antimalarial drug policy change

Case Study 3: National-Level Scale-up of malaria interventions

References Alexander K. Rowe, Faustin Onikpo, Marcel Lama, Dawn M. Osterholt, and Michael S. Deming Impact of a Malaria-Control Project in Benin That Included the Integrated Management of Childhood Illness Strategy. American Journal of Public Health, Published online ahead of print May 12, 2011, as /AJPH de Savigny, D and T Adam, eds Systems thinking for health systems strengthening. Geneva, Alliance for Health Policy and Systems Research and WHO. Craig, P, P Dieppe, et al Developing and evaluating complex interventions: new guidance. Medical Research Council. Galiani S, Gertler PJ, Schargrodsky E. Water for life: the impact of the privatization of water services on child mortality. J Polit Econ 2005;113: Gertler, Martinez, Premand, Rawlings and Vermeersch (2010) Impact Evaluation in Practice, Washington, DC: The World Bank. Habicht JP et al. (1999) Evaluation designs for adequacy, plausibility and probability of public health programme performance and impact. International Journal of Epidemiology, 28: Rossi P et al. (1999). Evaluation. A systematic Approach. Thousand Oaks: Sage Publications. Habicht, JP, CG Victora, et al Evaluation designs for adequacy, plausibility and probability of public health programme performance and impact. International Journal of Epidemiology 28:

Group work Please get into your project groups and select an evaluation design for your program (Experimental, Quasi-Experimental, Non- Experimental) Explain why you chose this design Discuss any strengths and weaknesses of this design as they relate to your program

General Questions to Be Answered by Each Group What study design will you use? What are the strengths and limitations of your evaluation design? How would you know if your program is effective? How will you address contextual or confounding factors?

MEASURE Evaluation is a MEASURE program project funded by the U.S. Agency for International Development (USAID) through Cooperative Agreement GHA-A and is implemented by the Carolina Population Center at the University of North Carolina at Chapel Hill, in partnership with Futures Group International, John Snow, Inc., ICF Macro, Management Sciences for Health, and Tulane University. Visit us online at