Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Discussant Comments Prepared by Carol.

Slides:



Advertisements
Similar presentations
Performance Assessment
Advertisements

Introduction to Monitoring and Evaluation
Standardized Scales.
Donald T. Simeon Caribbean Health Research Council
What You Will Learn From These Sessions
Supporting continuous improvement in the replication process Getting to Grips with Replication Seminar 3: Monitoring, evaluation & continuous improvement.
VALUE – ADDED 101 Ken Bernacki and Denise Brewster.
2006 International Symposium of Computer Assisted Language Learning,June 2-4, Beijing China Tutor feedback in online English language learning: tutor perceptions.
Just Because They Say It’s ‘Scientifically- based’ Doesn’t Mean It Will Work!
Facilitation with Fidelity Part 1: A Focus on Delivering Your Program with Fidelity South Carolina Campaign to Prevent Teen Pregnancy.
Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs David S. Cordray Vanderbilt University IES/NCER Summer Research.
Funding Opportunities at the Institute of Education Sciences Elizabeth R. Albro, Ph.D. Acting Commissioner, National Center for Education Research.
Funding Opportunities at the Institute of Education Sciences: Information for the Grants Administrator Elizabeth R. Albro, Ph.D. Acting Commissioner National.
Spending Public Money Wisely Scaling-Up Educational Interventions Barbara Schneider John A. Hannah University Distinguished Professor College of Education.
What We Know About Effective Professional Development: Implications for State MSPs Part 2 Iris R. Weiss June 11, 2008.
Striving for Quality Using continuous improvement strategies to increase program quality, implementation fidelity and durability Steve Goodman Director.
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Defining, Conceptualizing, and Measuring Fidelity of Implementation and Its Relationship to Outcomes in K–12 Curriculum Intervention Research Prepared.
Funding Opportunities at the Institute of Education Sciences Elizabeth R. Albro, Ph.D. Associate Commissioner Teaching and Learning Division National Center.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Achieved Relative Intervention Strength: Models and Methods Chris S. Hulleman David S. Cordray Presentation for the SREE Research Conference Washington,
Classroom Assessments Checklists, Rating Scales, and Rubrics
EDU 385 Education Assessment in the Classroom
Semester 2: Lecture 9 Analyzing Qualitative Data: Evaluation Research Prepared by: Dr. Lloyd Waller ©
Course Enhancement Module on Evidence-based Reading Instruction Collaboration for Effective Educator Development, Accountability and Reform H325A
Evaluating a Research Report
Assessment Practices That Lead to Student Learning Core Academy, Summer 2012.
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation Research Methods Meeting September 20-21, 2010.
Implementation and process evaluation: developing our approach Ann Lendrum University of Manchester Neil Humphrey University of Manchester Gemma Moss Institute.
KATEWINTEREVALUATION.com Education Research 101 A Beginner’s Guide for S STEM Principal Investigators.
Conceptualizing Intervention Fidelity: Implications for Measurement, Design, and Analysis Implementation: What to Consider At Different Stages in the Research.
GSSR Research Methodology and Methods of Social Inquiry socialinquiry.wordpress.com January 17, 2012 I. Mixed Methods Research II. Evaluation Research.
Student assessment AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
{ Principal Leadership Evaluation. The VAL-ED Vision… The construction of valid, reliable, unbiased, accurate, and useful reporting of results Summative.
ScWk 242 Course Overview and Review of ScWk 240 Concepts ScWk 242 Session 1 Slides.
1. Housekeeping Items June 8 th and 9 th put on calendar for 2 nd round of Iowa Core ***Shenandoah participants*** Module 6 training on March 24 th will.
Laying the Foundation for Scaling Up During Development.
CAUSAL INFERENCE Presented by: Dan Dowhower Alysia Cohen H 615 Friday, October 4, 2013.
IMPLEMENTATION QUALITY RESEARCH OF PREVENTION PROGRAMS IN CROATIA MIRANDA NOVAK University of Zagreb, Faculty of Education and Rehabilitation Sciences.
Evidence-based Education and the Culture of Special Education Chair: Jack States, Wing Institute Discussant: Teri Palmer, University of Oregon.
EDCI 696 Dr. D. Brown Presented by: Kim Bassa. Targeted Topics Analysis of dependent variables and different types of data Selecting the appropriate statistic.
The Conceptual Framework: What It Is and How It Works Linda Bradley, James Madison University Monica Minor, NCATE April 2008.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Consultant Advance Research Team. Outline UNDERSTANDING M&E DATA NEEDS PEOPLE, PARTNERSHIP AND PLANNING 1.Organizational structures with HIV M&E functions.
Developing a Framework In Support of a Community of Practice in ABI Jason Newberry, Research Director Tanya Darisi, Senior Researcher
+ Evidence Based Practice University of Utah Evidence-Based Treatment and Practice: New Opportunities to Bridge Clinical Research and Practice, Enhance.
Securing External Federal Funding Janice F. Almasi, Ph.D. Carol Lee Robertson Endowed Professor of Literacy University of Kentucky
An Expanded Model of Evidence-based Practice in Special Education Randy Keyworth Jack States Ronnie Detrich Wing Institute.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Symposium Chair: Chris S. Hulleman,
Updated Section 31a Information LITERACY, CAREER/COLLEGE READINESS, MTSS.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
Campbell Collaboration Colloquium 2014 "Better Evidence for a Better World" Why The U.S. Is So Bad At Knowledge Transfer and Implementation Randy Keyworth.
Cal-ABA 26th Annual Western Regional Conference What We Know About Sustaining Programs? Randy Keyworth Ronnie Detrich Jack States.
The Cause…or the “What” of What Works? David S. Cordray Vanderbilt University IES Research Conference Washington, DC June 16, 2006.
Funding Opportunities at the Institute of Education Sciences Elizabeth R. Albro, Ph.D. Teaching and Learning Division National Center for Education Research.
Internationalizing the Technology Education Curriculum Dr. Edward M. Reeve Professor Utah State University.
Stages of Research and Development
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 3. Session 9. Periodic data collection methods.
EDU 695Competitive Success/snaptutorial.com
EDU 695 Education for Service-- snaptutorial.com
EDU 695 Teaching Effectively-- snaptutorial.com
February 21-22, 2018.
Fidelity of Implementation in Scaling-up Highly Rated Science Curriculum Units for Diverse Populations Carol O’Donnell and Sharon Lynch The George Washington.
Some Further Considerations in Combining Single Case and Group Designs
Presentation transcript:

Progressing Toward a Shared Set of Methods and Standards for Developing and Using Measures of Implementation Fidelity Discussant Comments Prepared by Carol O’Donnell Institute of Education Sciences Society for Research on Educational Effectiveness Annual Meeting March 5, 2010

Thank you to the authors for inviting me to participate in this symposium and for providing their papers in advance. I commend the authors on their innovative work and for their work in guiding the field in this unexplored and new territory. My remarks that follow are my own and do not necessarily represent the views of the US Department of Education and the Institute of Education Sciences. Introduction

Papers’ Common Themes: What problems exist? Michael’s “black box”—if we want to determine effectiveness of an intervention, we need to define the treatment and its counterfactual. If “it” works, what is “it”? Developers often fail to identify the “critical components” of an intervention. Researchers often fail to measure whether components were delivered as intended. Did they have a greater influence on the DV than the comparison? How does implementation differ in the treatment and comparison group?

Common Themes: What problems exist? Most fidelity studies are correlational--not causal. Analyses show the size of effects obtained when interventions are implemented under classroom conditions are smaller than size of effects obtained during laboratory trials. The same has been found in meta-analyses comparing efficacy (under ideal conditions) & effectiveness studies (routine conditions). Lipsey, 1999; Petrosino & Soydan, 2005; Weisz, Weiss & Donenberg, 1992

Common Themes: Summary of Problems A.Lack of an attempt to define the intervention a priori (using a logic or change model). B.Lack of consensus on how to measure fidelity. C.Lack of attempts to use fidelity to analyze outcomes, especially if multi-dimensional. D.Lack of consistency on when and how to assess and report fidelity.

Solutions to Problems A-D (which correspond somewhat with the authors’ 5 step procedure for fidelity assessment)

Solution A: Define the intervention a priori Developers need to make the critical components explicit to users, distinguishing them between structure and process. Nice distinction between change and logic model, and between a descriptive model of implementation (what transpired as the intervention was put in place), vs. an a priori model which has explicit expectations about implementation of program components. Michael’s “intervention as implemented vs intervention as designed.” Anne noted challenges when intervention is unscripted. Not clear how we know the components are critical. Exploratory work should precede development.

Catherine - It’s important to distinguish “fidelity to process” from “global class quality” (context variables such as good classroom management). “Global class quality” variables are only fidelity variables if they are promoted by the intervention; otherwise, they are a separate construct from fidelity. Such variables may impact the effects of fidelity on outcomes (mediation moderation model); therefore, they are important to measure. Solution B: Measure fidelity

When possible, include fidelity measures in meta-analyses of efficacy studies. For sets of studies with broad overarching goals (such as PCER), consider whether there is a global fidelity model and measure, despite multiple interventions.

Solution B: Measure fidelity I commend the authors who argue for a standard for fidelity measures in education program evaluation. What can we do promote standardization? –IES requires researchers to conceptualize fidelity during development, assess fidelity during efficacy and scale-up evaluations, and use fidelity data in analyses (RFA now cites Hulleman & Cordray, 2009). –Teacher education programs and professional development should help teachers to be good consumers of curriculum materials and understand the role fidelity plays in its selection, implementation, and evaluation. –Psychometricians should be involved in the development of fidelity measures—examining their validity and reliability.

Solution C: Analyze fidelity data Michael mentioned the limitations of the intent-to-treat experimental model for explaining effects. Analysis of fidelity data varies greatly: –Descriptive - frequency or percentage of fidelity –Associative – simple correlations –Predictive – explaining variance in outcomes –Causal – RCT of teachers to high and low fidelity (rare) –Impact – fidelity as 3 rd moderator or mediator –Adjusting Outcomes – achieved relative strength vs comparing effect sizes of fidelity vs infidelity Analysis of fidelity data often disparate—what does it all mean? Need a more complete fidelity assessment to better understand construct validity and generalizability.

Solution C: Analyze fidelity data Achieved Fidelity vs Achieved Relative Strength Chris – intervention may fail to have effects because it does not differ from control on core components.

Solution D: Standardize when & how to assess fidelity Monitor fidelity during efficacy studies to gain confidence that outcomes are due to intervention (internal validity), and, Michael pointed out, to distinguish between implementation failure vs program failure. Determine if results under a specific structure replicate under other organizational structures (Chris’ operational model). Anne discussed fidelity’s impact on ongoing program development. Fidelity results should inform revisions. Also, decide what components of the intervention can be modified to better suit specific conditions without affecting the efficacy of the intervention.

Solution D: Standardize when & how Understand the implementation conditions, tools, and processes needed to reproduce positive effects under routine practice on a large scale (i.e., if fidelity is moving target, generalizability of scale-up research may be imperiled). Decide what training is required to provide those who will deliver the intervention with the knowledge and skills to deliver the intervention as intended. Decide if training is a part of “it”.

Questions to Consider

What role does the counterfactual play in the conception of the change and logic models? Is the structure and process framework similar to the change (constructs/process) v. logic model (structural)? Should fidelity measures for the comparison group (process only?) be different than fidelity measures for the treatment group (structure and process)? Positive infidelity – was this a result of “global class quality”? Is it really “infidelity” (which is NOT implementing a critical component) or was it just supplementing the curriculum (which has been shown to enhance outcomes (as long as fidelity is high). Questions to Consider: Define intervention

Is one fidelity measure enough? (The need for multiple measures enriches fidelity data, but complicates the model conceptually and structurally when measures are integrated. Multiple measures may inflate standard errors of parameter estimators. Need for parsimony.) Are scores on items additive? Subjects receiving the same fidelity score may have different implementation profiles (and ordinal scores are not interval). Can fidelity measures be universal, or program-specific? Standardize methods—not measures. Questions to Consider: Measure fidelity

Questions to Consider: Analyze fidelity data What are the dangers of removing low fidelity implementers from the sample, or creating bivariate median split between high and low fidelity users (which loses continuous data)? If an intervention is implemented at the classroom level, what is the unit of analysis—think about participant responsiveness (student fidelity, teacher fidelity, school fidelity)? Catherine pointed out that child fidelity was often ignored. What role does HLM play?

How do researchers encourage fidelity when creativity, variability, and local adaptations are encouraged? How do we distinguish between fidelity and adaptation? Are the intervention effects established under careful monitoring financially feasible? Questions to Consider: Standardize when & how

Bottom Line: Can the intervention be implemented with adequate fidelity under conditions of routine practice and yield positive results? Source: O’Donnell, 2008a

Thank you.