PPA 502 – Program Evaluation Lecture 3c – Strategies for Impact Assessment.

Slides:



Advertisements
Similar presentations
PhD Research Seminar Series: Valid Research Designs
Advertisements

Chapter 22 Evaluating a Research Report Gay, Mills, and Airasian
GROUP-LEVEL DESIGNS Chapter 9.
Experimental Research Designs
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
Non-Experimental designs: Developmental designs & Small-N designs
The Methods of Social Psychology
Non-Experimental designs: Developmental designs & Small-N designs
Psych 231: Research Methods in Psychology
Lecture 5 Data Coding and Experimental Research Methods.
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
TOOLS OF POSITIVE ANALYSIS
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Types of Evaluation.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Experimental Research
Formulating the research design
Basic Principles of Research Design
Experimental Design The Gold Standard?.
Learning Objective Chapter 8 Primary Data Collection: Experimentation CHAPTER eight Primary Data Collection: Experimentation Copyright © 2000 by John Wiley.
Research Design for Quantitative Studies
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Primary Data Collection: Experimentation CHAPTER eight.
Chapter 3 The Research Design. Research Design A research design is a plan of action for executing a research project, specifying The theory to be tested.
Quantitative Research Designs
Day 6: Non-Experimental & Experimental Design
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Epidemiology The Basics Only… Adapted with permission from a class presentation developed by Dr. Charles Lynch – University of Iowa, Iowa City.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Learning Objectives Copyright © 2004 John Wiley & Sons,Inc Primary Data Collection: Experimentation CHAPTER Seven.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Evaluating a Research Report
 Internal Validity  Construct Validity  External Validity * In the context of a research study, i.e., not measurement validity.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Primary Data Collection: Experimentation CHAPTER eight.
Techniques of research control: -Extraneous variables (confounding) are: The variables which could have an unwanted effect on the dependent variable under.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 8 Planning a Nursing Study.
IDEV 624 – Monitoring and Evaluation Introduction to Process Monitoring Payson Center for International Development and Technology Transfer Tulane University.
Chapter Four Experimental & Quasi-experimental Designs.
URBDP 591 I Lecture 3: Research Process Objectives What are the major steps in the research process? What is an operational definition of variables? What.
PPA 502 – Program Evaluation Lecture 2c – Process Evaluation.
Quantitative and Qualitative Approaches
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Begin at the Beginning introduction to evaluation Begin at the Beginning introduction to evaluation.
Reading and Evaluating Research Method. Essential question to ask about the Method: “Is the operationalization of the hypothesis valid? Sections: Section.
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Experimental & Quasi-Experimental Designs Dr. Guerette.
EXPERIMENTS AND EXPERIMENTAL DESIGN
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Designing New Programs Design & Chronological Perspectives (Presentation of Berk & Rossi’s Thinking About Program Evaluation, Sage Press, 1990)
Evaluation Research Dr. Guerette. Introduction Evaluation Research – Evaluation Research – The purpose is to evaluate the impact of policies The purpose.
to become a critical consumer of information.
EXPERIMENTS Lecture 5. Administrative STATA Course Mailing List Info:   No subject  In body.
Design of Clinical Research Studies ASAP Session by: Robert McCarter, ScD Dir. Biostatistics and Informatics, CNMC
Chapter Nine Primary Data Collection: Experimentation and
Experiments.  Labs (update and questions)  STATA Introduction  Intro to Experiments and Experimental Design 2.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
SOCI 4466 PROGRAM & POLICY EVALUATION LECTURE #8 1. Evaluation projects 2. Take-home final 3. Questions?
How do you know your product “works”? And what does it mean for a product to “work”?
Experimental Design. Talking While Driving: Are Cell Phone Conversations Different from Passenger Conversations? Research Problems Talking on a cellphone.
CRITICALLY APPRAISING EVIDENCE Lisa Broughton, PhD, RN, CCRN.
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
MONITORING AND EVALUATION IN TB/HIV PROGRAMS
Presentation transcript:

PPA 502 – Program Evaluation Lecture 3c – Strategies for Impact Assessment

Introduction  The ultimate purpose of a social program is to ameliorate some social problem or improve some social condition. If the program theory is sound and the program plan well implemented, those social benefits are expected to follow. Rarely are those benefits assured, however. Practical and conceptual shortcomings combined with the intractable nature of many social problems all too easily undermine the effectiveness of social programs.

Introduction  A general principle applies: The more rigorous the research design, the more plausible the resulting estimate of intervention effects.  The design of impact evaluations faces two competing pressures: –Evaluations should be undertaken with sufficient rigor that relatively firm conclusions can be reached. –Practical considerations of time, money, cooperation, and protection of participants limit the design options and methodological procedures that can be employed.

Introduction  Evaluators assess the effects of social programs by: –Comparing information about outcomes for participants and nonparticipants, –Making repeated measurements on participants before and after intervention, –Or other methods that attempt to achieve the equivalent of such comparisons.  The basic aim of impact assessment is to produce an estimate of the net effects of an intervention.

Introduction  Impact assessment is relevant at many stages of the process. –Pilot demonstrations to estimate whether a proposed program will work. –Program design to test the most effective ways to develop and integrate the various program elements. –Program initiation to test the efficacy of the program at a limited number of sites. –Program modification to test the effects of the changes. –Program continuation to test for sunset legislation, funding renewal, or program defense.

Key Concepts in Impact Assessment  The experimental model. –The optimal way to assess impact is a randomized field experiment. Random assignment Treatment and control groups Net outcome assessment.  Prerequisites for assessing impact. –Program’s objectives must be well-articulated to make it possible to specify credible measures of the expected outcomes. –The interventions must be sufficiently well- implemented that there is no question that critical elements have been delivered to appropriate targets.

Key Concepts in Impact Assessment  Linking interventions to outcomes. –Establishing impact essentially amounts to establishing causality. –Most causal relationships in social science expressed as probabilities. –Conditions limited causality External conditions and causes. Biased selection. Other social programs with similar targets.

Key Concepts in Impact Assessment  “Perfect” versus “good enough” impact assessments. –Intervention and target may not allow perfect design. –Time and resource constraints. –Importance often determines rigor. –Review design options to determine most appropriate.

Key Concepts in Impact Assessment  Gross versus net outcomes.

Extraneous Confounding Factors  Uncontrolled selection. –Preexisting differences between treatment and control groups. –Self-selection. –Program location and access. –Deselection processes (attrition bias).  Endogenous change. –Secular drift. –Interfering events. –Maturational trends.

Design Effects  Stochastic effects. –Significance (Type I error). –Power (Type II error). –The key is finding the proper balance between the two.  Measurement reliability. –Does the measure produce the same results repeatedly? –Unreliability dilutes and obscures real differences. –Reproducibility should not fall below 75 or 80%.

Design effects  Measurement validity. –Does the instrument measure what it is intended to measure? –Criteria. Consistency with usage. Consistency with alternative measures. Internal consistency. Consequential predictability.

Design Effects  Choice of outcome measures. –A critical measurement problem in evaluations is that of selecting the best measures for assessing outcomes. Conceptualization. Reliability. Feasibility. Proxy.  The Hawthorne Effect. –Delivery affected by context.

Design Effects  Missing information. –Missing information is generally not randomly distributed. –Often must be supplemented by alternative survey items, unobtrusive measures, or estimates.

Design Effects  Sample design effects. –Must select an unbiased sample of the universe of interest. Select a relevant sensible universe. Design a means of selecting an unbiased sample (random). Implement sample design with fidelity.  Minimizing design effects. –Planning. –Pretesting. –Sampling.

Design Strategies for Isolating the Effects of Extraneous Factors  Randomized controls.  Regression-discontinuity controls.  Matched construct controls.  Statistically equated controls.  Reflexive controls.  Repeated measures reflexive controls.  Time-series reflexive controls.  Generic controls.

Design Strategies for Isolating the Effects of Extraneous Factors  Full- versus partial-coverage programs. –Full coverage means absence of control group. Must use reflexive controls.

A Catalog of Impact Assessment Designs Research DesignIntervention AssignmentTypes of Controls UsedData Collection I. Designs for partial coverage programs A. Randomized or true experimentsRandom assignment controlled by researcher Experimental and control groups randomly selected Minimum data needed are after- intervention measures B. Quasi-experiments 1. Regression-discontinuityNonrandom but fixed and none to researcher Selected targets compared to unselected targets, holding selection constant Typically consists of multiple before- and after-intervention outcome measures. 2. Matched controlsNonrandom and unknownIntervention group matched with controls selected by researcher Typically consists of before- and after- intervention measures 3. Statistically equated controlsNonrandom and often nonuniformExposed and unexposed targets compared by means of statistical controls Before-and-after or after-only intervention outcome measures and control variables 4. Generic controlsNonrandomExposed target compared with outcome measures available for general population After-intervention outcome measures on targets plus publicly available norms of outcome levels in the general population. II. Designs for full-coverage programs A. Simple before-and-after studiesNonrandom and uniformTargets measured before and after intervention Outcome measured on exposed targets before and after intervention. B. Cross-section studies for nonuniform programs Nonrandom and nonuniformTargets differentially exposed to intervention compared with statistical controls After-intervention outcome measures and control variables C. Panel studies: Several repeated measures for nonuniform programs Nonrandom and nonuniformTargets measured before, during, and after intervention Repeated measures of exposure to intervention and of outcome D. Time series: Many repeated measures Nonrandom and nonuniformLarge aggregates cmpared before and after intervention Many repeated before- and after- intervention outcome measures on large aggregates

Judgmental Approaches to Impact Assessment  Connoisseurial impact assessments.  Administrator impact assessments.  Participant’s assessments.  The use of judgmental assessments. –Limited funds. –No preintervention measures. –Full-coverage, uniform programs.

Inference Validity Issues in Impact Assessment  Reproducibility (can a researcher using the same design in the same setting achieve the same results?). –Power of design. –Fidelity of implementation. –Appropriateness of statistical models.

Inference Validity Issues in Impact Assessment  Generalizability (the applicability of the findings to similar situations that were not studied.). –Unbiased sample. –Faithful representation of actual program. –Common settings.  Stress reproducibility first through several iterations, then focus on generalizability.  Pooling evaluations: meta-analysis.