Which Evaluation Designs Are Right for Your State?

Slides:



Advertisements
Similar presentations
Program Evaluation: What is it?
Advertisements

Evaluation Capacity Building Identifying and Addressing the Fields Needs.
Program Evaluation It’s Not Just for OMB Anymore….
PAI786: Urban Policy Class 2: Evaluating Social Programs.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
1 Presented By: Dr. Jacob BenusDr. Wayne Vroman Project DirectorPrincipal Investigator July 11-13, 2005 The Reemployment Eligibility Assessment (REA) Study.
Fundamentals of Evaluation for Public Health Programs ROBERT FOLEY, M.ED. NIHB TRIBAL PUBLIC HEALTH SUMMIT MARCH 31,
Designing a Random Assignment Social Experiment In the U.K.; The Employment Retention and Advancement Demonstration (ERA)
Early Childhood Outcomes Center1 Using Data for Program Improvement Christina Kasprzak, NECTAC/ECO Ann Bailey, NCRRC July 2010.
Quantitative and Qualitative Approaches
Rigorous Quasi-Experimental Evaluations: Design Considerations Sung-Woo Cho, Ph.D. June 11, 2015 Success from the Start: Round 4 Convening US Department.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
“A Truthful Evaluation Of Yourself Gives Feedback For Growth and Success” Brenda Johnson Padgett Brenda Johnson Padgett.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
The Disability Employment Initiative (DEI): Impact Evaluation Design October 21, 2015 Sung-Woo Cho, Ph.D.
1 General Elements in Evaluation Research. 2 Types of Evaluations.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Welcome to Workforce 3 One U.S. Department of Labor Employment and Training Administration Webinar Date: April 6 th, 2015 Presented by: U.S. Department.
Chapter 11 Experimental Designs PowerPoint presentation developed by: Sarah E. Bledsoe & E. Roberto Orellana.
Chapter 6 Selecting a Design. Research Design The overall approach to the study that details all the major components describing how the research will.
Evaluating the Quality and Impact of Community Benefit Programs
Technical Business Consultancy Project
Issues in Evaluating Educational Research
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Part 1: Program Evaluation: What is it?
DATA COLLECTION METHODS IN NURSING RESEARCH
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 3. Session 9. Periodic data collection methods.
The assessment process For Administrative units
Gender-Sensitive Monitoring and Evaluation
Gender-Sensitive Monitoring and Evaluation
Measuring Results and Impact Evaluation: From Promises into Evidence
Technical Assistance on Evaluating SDGs: Leave No One Behind
Child Outcomes Summary Process April 26, 2017
Public School Monitoring Roadmap
Right-sized Evaluation
Fundamentals of Monitoring and Evaluation
MUHC Innovation Model.
Clinical Studies Continuum
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Chapter 2 Sociological Research Methods
Program Evaluation Essentials-- Part 2
Reading Research Papers-A Basic Guide to Critical Analysis
America’s Promise Evaluation What is it and what should you expect?
Chapter Eight: Quantitative Methods
© 2012 The McGraw-Hill Companies, Inc.
It’s Not Just for OMB Anymore…
Ian D. Rotherham Professor of Environmental Geography,
Integrating Outcomes Learning Community Call February 8, 2012
Development Impact Evaluation in Finance and Private Sector
Using outcomes data for program improvement
ABAB Design Ethical considerations
Implementation Challenges
Using Data for Program Improvement
REAL (working with Singizi) Presentation to NSA 3rd/4th August
February 19, 2019 An Evaluation Readiness Assessment— Prioritizing Efforts to Measure Effectiveness Evaluations and Research Hub Flash Webinar Series.
Class 2: Evaluating Social Programs
Class 2: Evaluating Social Programs
Using Data for Program Improvement
Research Design Quantitative.
Introduction to the design (and analysis) of experiments
Monitoring and Evaluating FGM/C abandonment programs
Using the Child and Family Outcomes Analysis Tools
Evaluation and Research to Increase Evidence for Workforce Programs
How does evaluation help my state and where do we start?
RESEA Evaluation Technical Assistance
Procuring and Selecting an Independent Evaluator
Run of Show Goals of RESEA EvalTA (Cycles of Learning and Doing) 0:51
Run of Show Today’s Objectives 4:53 Participant Polls 8:36
Rate your experience level producing the Annual Statewide Performance Report Narrative Master: I am well versed in all aspects of the WIOA Statewide.
Presentation transcript:

Which Evaluation Designs Are Right for Your State? May 22, 2019 Which Evaluation Designs Are Right for Your State? Reemployment Service and Eligibility Assessment (RESEA) Evaluation Technical Assistance (EvalTA)

Gloria Salas-Kos Senior Program Analyst and RESEA Evaluation TA Coordinator Office of Policy Development and Research, ETA, U.S. DOL

Evaluating RESEA: How Does it Help My State and Where Do We Start? Our previous webinar on May 2, 2019 provides an overview of the following topics: Program evaluation and its benefits for RESEA programs; Tools to help you form learning goals and think about your potential evaluation efforts; and Key evaluation concepts about research questions and evaluation designs.

Describe implementation and process studies and what you can learn from them; Discuss the concept of “impact” and how to measure it using control or comparison groups; and Provide an introduction to different impact evaluation designs. Today’s presentation provides mid-level of complexity of evaluation methods, terms, and approaches to consider.

Evaluation Design Clarifies the program goals and strategies, Identifies relevant and useful research questions, Determines the approach type of evaluation, Identifies relevant data sources, and Describes data analysis plans that allow for valid conclusions to be made. ~ Designing Evaluations, GAO Applied Research and Methods, 2012

Different Types of Evaluations Needs assessment Literature reviews Pre-evaluation Assessment Implementation studies Process studies Descriptive Random assignment Quasi-experimental design (QED) Impact

Questions for Your RESEA Evaluations How to Answer It What really happens in our RESEA program? Implementation and Process Study Does our RESEA program improve outcomes? Would an alternative strategy yield even better outcomes? Impact Study Strongest Approach: Random assignment Quasi-experimental designs (QEDs) How are we going to do this? DOL and Abt are here to help!

Andrew Clarkwest Tresa Kappil Senior Associate Senior Analyst Abt Associates Tresa Kappil Senior Analyst Abt Associates

Questions for Your RESEA Evaluations How to Answer It What really happens in our RESEA program? Implementation and Process Study Does our RESEA program improve outcomes? Would some alternative strategy yield even better outcomes? Impact Study Strongest Approach: Random Assignment Quasi-experimental designs (QEDs) How are we going to do this? DOL and Abt are here to help!

What are Implementation and Process Evaluations? Identify how closely program implementation adheres to plans, Including the variations that exist across locations (Workforce Development Boards and Regions, American Job Centers, etc.) Explain “what is happening and why” to help understand what claimants’ experiences are and to support quick feedback in areas to improve quality or efficiency.

Implementation and Process Studies Can... Provide feedback to program managers, and accountability to program sponsors and the public Highlight best practices and identify areas that can be strengthened Inform future program development or support replication Help interpret impact study findings

What do Implementation and Process Studies involve? Identify research questions of interest Systematically collect data through: Interviews with staff, claimants, and stakeholders, Focus groups with participants Review of program documents Observation of program activities Administrative or survey data Analyze themes in data using qualitative research methods Summarize and communicate findings

Implementation and Process Study Limitations Researchers may have to make judgments about what certain types qualitative and/or subjective data are telling them Cannot tell you whether your intervention causes changes in claimant outcomes ~Werner, A. (2012). Guide to Implementation Research. Urban Institute Press

Questions for Your RESEA Evaluations How to Answer It What really happens in our RESEA program? Implementation and Process Study Does our RESEA program improve outcomes? Would some alternative strategy yield even better outcomes? Impact Study Strongest Approach: Random assignment Quasi-experimental designs (QED) How are we going to do this? DOL and Abt are here to help!

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What is involved in executing the evaluation design?

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What is involved in executing the evaluation design?

What is “Impact”? Treatment Outcomes w/ intervention vs. Counterfactual Outcomes w/o intervention Holding all else equal

What is “Impact”? Treatment Outcomes w/ intervention vs. Counterfactual Outcomes w/o intervention Holding all else equal The impact of the RESEA intervention is defined as … … the difference between outcomes with the RESEA intervention vs. outcomes w/o the RESEA intervention

What is “Impact”? Treatment Outcomes w/ intervention vs. Counterfactual Outcomes w/o intervention Holding all else equal The impact of the RESEA intervention is defined as … … the difference between outcomes with the RESEA intervention vs. outcomes w/o the RESEA intervention … holding all else equal

Three Types of Impact Research Questions Does my RESEA program “make a difference”? Does a component of my RESEA program “make a difference”? Would an alternative strategy “make a bigger difference”?

Three Types of Impact Research Questions Does my RESEA program “make a difference”? Does a component of my RESEA program “make a difference”? Would an alternative strategy “make a bigger difference”? Are outcomes better than they would be if the claimant had not been assigned to the program? How large is the difference?

Three Types of Impact Research Questions Does my RESEA program “make a difference”? Does a component of my RESEA program “make a difference”? Would an alternative strategy “make a bigger difference”? E.g, Do intensive reemployment services received by claimants contribute to RESEA program impacts? How much of a difference do they make?

Three Types of Impact Research Questions Does my RESEA program “make a difference”? Does a component of my RESEA program “make a difference”? Would an alternative strategy “make a bigger difference”? E.g., Would using different needs assessments change employment outcomes? Which way (better or worse)? How large a difference?

How is Impact Different from Outcome?

How is Impact Different from Outcome?

How is Impact Different from Outcome?

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What practical concerns do I need to address to execute an evaluation of this type?

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What practical concerns do I need to address to execute an evaluation of this type?

Can the Design Give Me Valid Results? Treatment Outcomes w/ intervention vs. Counterfactual Outcomes w/o intervention Holding all else equal We only see what happens under one condition, not the other, so the challenge is: How do we know what would have happened otherwise? Impact evaluations aim to solve that challenge

Consideration 1: Is the Counterfactual Credible? We want our study to estimate “impact” Do the results reflect only the effects of the intervention, not other factors? To estimate impact we need a good counterfactual A comparison source that shows us what outcomes our treatment group would have had if they had not received the treatment. What might that be?

Is the Counterfactual Credible? An example Can a state get an accurate estimate of the impact of its RESEA program by comparing: The outcomes of RESEA claimants to The outcomes of claimants not selected for RESEA?

Is the Counterfactual Credible? An Example Example: Comparing RESEA claimants to claimants not selected for RESEA We want these two bars to be the same

Is the Counterfactual Credible? An Example Example: Comparing RESEA claimants to claimants not selected for RESEA We want these two bars to be the same

Is the Counterfactual Credible? An Example Example: Comparing RESEA claimants to claimants not selected for RESEA True Impact Estimated Impact

Is the Counterfactual Credible? An example Can a state get an accurate estimate of the impact of its RESEA program by comparing: The outcomes of RESEA claimants to The outcomes of claimants not selected for RESEA? NO! Because all else is not equal

Consideration 1: Is the Counterfactual Credible? CLEAR Through CLEAR, DOL has established standards for whether a study has established a credible counterfactual (https://clear.dol.gov/) CLEAR’s ratings indicate how confident we can be that a study’s findings reflect the impact of an intervention, rather than something else You will want your impact evaluations to meet CLEAR standards CLEAR has standards for different designs Random assignment (experimental) Other designs (quasi-experimental)

How Random Assignment Creates a Strong Counterfactual For each person, flip a coin Which is random …

How Random Assignment Creates a Strong Counterfactual For each person, flip a coin Which is random … Heads … Get the intervention Only (systematic) difference between the groups is the intervention Tails … Do not get the intervention i.e., random assignment “holds all else equal”

How Random Assignment Creates a Strong Counterfactual For each person, flip a coin Which is random … Heads … Get the intervention Only (systematic) difference between the groups is the intervention Tails … Do not get the intervention So any difference in outcomes must be due to the program

Question and Strategies for Your RESEA Evaluation - QEDs Strategy What really happens in our RESEA program? Implementation and Process Study Does our RESEA program improve outcomes? Would some alternative strategy yield even better outcomes? Impact Study Strongest Approach: Random Assignment Quasi-experimental designs (QED) How are we going to do this? DOL and Abt are here to help!

Other Approaches to Estimating the Counterfactual -- Matching Find people who did not receive the intervention who look like those who did in all observable ways Blue: Selected for Intervention Red: Not Selected for Intervention

Other Approaches to Estimating the Counterfactual -- Matching Find people who did not receive the intervention who look like those who did in all observable ways Blue: Selected for Intervention Red: Not Selected for Intervention

Other Approaches to Estimating the Counterfactual -- Matching Find people who did not receive the intervention who look like those who did in all observable ways Blue: Received the Intervention Red: Did Not Receive the Intervention

Other Approaches to Estimating the Counterfactual – Matching Limitations: There may be few or no comparison claimants that really are similar on observables to intervention claimants There may be large differences on unobserved characteristics that we cannot account for

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What is involved in executing the design?

Consideration 2: How Many Claimants Need to Be in the Study? Example of a very small sample Sample of 3 comparison group members and 3 treatment group members. Suppose 66.7% (2 of 3) of comparison group members get a job 33.3% (1 of 3) of treatment group members get a job Did the program have an impact of -33%? Small sample sizes lead to inconclusive results But how large of a sample do you need?

Consideration 2: How Many Claimants Need to Be in the Study? Required sample sizes are often surprisingly large Sample sizes needed depends on: Outcomes of interest How large the difference is between the RESEA intervention and the counterfactual intervention Study design

Consideration 2: How Many Claimants Need to Be in the Study? Likely Minimum Sample Sizes, by Outcome Intermediate outcomes: Hundreds of claimants UI outcomes: Thousands of claimants Labor market outcomes: Tens of thousands of claimants Statute requires evidence of impact on UI outcomes and labor market outcomes Sample sizes for QEDs are larger, sometime much larger than for random assignment designs

Questions about Impact Design What do we mean by “impact”? What should I consider when choosing a method to estimate impact? Credibility: Can the design give me valid results? Sample Size: How large of a sample of claimants do we need to detect differences in the outcomes of interest? Logistics: What is involved in executing the evaluation design?

Consideration 3: What Logistical Requirements Do I Need to Consider? Possible changes to program implementation Time required to conduct evaluation Technical expertise IT and data resources

Consideration 3: What Logistical Requirements Do I Need to Consider? Possible changes to program implementation Time required to conduct evaluation Technical expertise IT and data resources More intense when testing alternative services May need to train staff on new procedures RCTs may require more oversight if service provision has to be monitored

Consideration 3: What Logistical Requirements Do I Need to Consider? Possible changes to program implementation Time required to conduct evaluation Technical expertise IT and data resources If you can use existing data it will take less time Possible only with QEDs If you have to gather new data, it will require more time RCTs Any tests of new interventions

Consideration 3: What Logistical Requirements Do I Need to Consider? Possible changes to program implementation Time required to conduct evaluation Technical expertise IT and data resources Substantial expertise is required to successfully implement any high quality impact evaluation QEDs are more technically complex than random assignment to carry out

Consideration 3: What Logistical Requirements Do I Need to Consider? Possible changes to program implementation Time required to conduct evaluation Technical expertise IT and data resources Identify sources of all needed measures, and assess data quality QEDs: Require a richer set of background characteristics Random assignment: Incorporate random assignment into RESEA selection algorithms

Question for Your RESEA Evaluations How to Answer It What really happens in our RESEA program? Implementation and Process Study Does our RESEA program improve outcomes? Would some alternative strategy yield even better outcomes? Impact Study Randomized Trial (the leading design) Other designs How are we going to do this? DOL and Abt are here to help!

Stay Tuned for more Evaluation Technical Assistance (EvalTA)! Look for the RESEA webinar series schedule with more evaluation topics Expect an RESEA-specific evaluation toolkit as well as a series of briefs Ask for one-on-one customized EvalTA Walks you systematically through needed planning steps and provides feedback on ideas Contact RESEA@abtassoc.com for more information

Procuring and Selecting an Independent Evaluator What Evaluation Details Do I Need for a Plan and How Long Will It Take? Week of June 17-21, 2019 Procuring and Selecting an Independent Evaluator Week of July 15-19, 2019 Using CLEAR – A Demonstration Week of August 5-9, 2019

U. S. DOL Guidance on RESEA and Evaluations Lawrence Burns Reemployment Coordinator Office of Unemployment Insurance Megan Lizik Senior Evaluation Specialist and Project Officer for RESEA Evaluation Chief Evaluation Office Wayne Gordon Director Division of Research and Evaluation

RESEA EvalTA Inbox Larry Burns Megan Lizik Gloria Salas-Kos Reemployment Coordinator U.S. DOL – Office of Unemployment Insurance Burns.Lawrence@dol.gov 202-693-3141 Megan Lizik Senior Evaluation Specialist and Project Officer U.S. DOL – Chief Evaluation Office Lizik.Megan@dol.gov Gloria Salas-Kos Evaluation Technical Assistance Coordinator U.S. DOL – Office of Policy Development and Research Salas-Kos.Gloria@dol.gov Andrew Clarkwest Senior Associate Abt Associates Andrew_Clarkwest@abtassoc.com 301.347.5065 RESEA EvalTA Inbox RESEA@abtassoc.com Tresa Kappil Senior Analyst Abt Associates Tresa_Kappil@abtassoc.com 301.347.5923