Standards of Evidence for Prevention Programs Brian R. Flay, D.Phil. Distinguished Professor Public Health and Psychology University of Illinois at Chicago.

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

Evaluation Research Pierre-Auguste Renoir: Le Grenouillere, 1869.
Chapter 3 Flashcards. obligation of an individual to other individuals based on a social or legal contract to justify his or her actions; the processes.
Experimental Research Designs
Center for the Study and Prevention of Violence University of Colorado Boulder
Program Evaluation. Lecture Overview  Program evaluation and program development  Logic of program evaluation (Program theory)  Four-Step Model  Comprehensive.
Assessing Program Impact Chapter 8. Impact assessments answer… Does a program really work? Does a program produce desired effects over and above what.
+ Evidence Based Practice University of Utah Presented by Will Backner December 2009 Training School Psychologists to be Experts in Evidence Based Practices.
Evaluation is a professional and ethical responsibility and is a core part of PHN professional practice Commitment to evaluation helps build the PHN intelligence.
Program Evaluation Spero Manson PhD
1 Design Approaches to Causal Inference Statistical mediation analysis answers the following question, “How does a researcher use measures of the hypothetical.
Types of Evaluation.
Studying treatment of suicidal ideation & attempts: Designs, Statistical Analysis, and Methodological Considerations Jill M. Harkavy-Friedman, Ph.D.
Experimental Design The Gold Standard?.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Chapter 4 Principles of Quantitative Research. Answering Questions  Quantitative Research attempts to answer questions by ascribing importance (significance)
INTRODUCTION TO OPERATIONS RESEARCH Niranjan Saggurti, PhD Senior Researcher Scientific Development Workshop Operations Research on GBV/HIV AIDS 2014,
STUDY PLANNING & DESIGN TO ENHANCE TRANSLATION OF HEALTH BEHAVIOR RESEARCH Lisa Klesges, Russell Glasgow, Paul Estabrooks, David Dzewaltowski, Sheana Bull.
Research Design for Quantitative Studies
RESEARCH A systematic quest for undiscovered truth A way of thinking
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 18 Mixed Methods and Other Special Types of Research.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 12 Undertaking Research for Specific Purposes.
Standards of Evidence for Substance Abuse (SU) Prevention Brian R. Flay, D.Phil. Distinguished Professor Public Health and Psychology University of Illinois.
Course on Data Analysis and Interpretation P Presented by B. Unmar Sponsored by GGSU PART 2 Date: 5 July
Chapter 3 Research Methods Used to Study Child Behavior Disorders.
Evaluating a Research Report
Implementation and process evaluation: developing our approach Ann Lendrum University of Manchester Neil Humphrey University of Manchester Gemma Moss Institute.
Copyright ©2008 by Pearson Education, Inc. Pearson Prentice Hall Upper Saddle River, NJ Foundations of Nursing Research, 5e By Rose Marie Nieswiadomy.
UIC University of Illinois at Chicago Health Research and Policy Centers “Phases of Research and Development for Health Promotion Programs” AAHB Annual.
STANDARDS OF EVIDENCE FOR INFORMING DECISIONS ON CHOOSING AMONG ALTERNATIVE APPROACHES TO PROVIDING RH/FP SERVICES Ian Askew, Population Council July 30,
Secondary Translation: Completing the process to Improving Health Daniel E. Ford, MD, MPH Vice Dean Johns Hopkins School of Medicine Introduction to Clinical.
Systematic reviews to support public policy: An overview Jeff Valentine University of Louisville AfrEA – NONIE – 3ie Cairo.
Laying the Foundation for Scaling Up During Development.
1 Experimental Research Cause + Effect Manipulation Control.
Methodology Matters: Doing Research in the Behavioral and Social Sciences ICS 205 Ha Nguyen Chad Ata.
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
Standards of Evidence for Program Efficacy, Effectiveness and Readiness for Dissemination Society for Prevention Research (SPR) Brian R. Flay, D.Phil.
Chapter 6 Training Evaluation
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
Evidence-based Education and the Culture of Special Education Chair: Jack States, Wing Institute Discussant: Teri Palmer, University of Oregon.
+ IDENTIFYING AND IMPLEMENTING EDUCATIONAL PRACTICES SUPPORTED BY RIGOROUS EVIDENCE: A USER FRIENDLY GUIDE Presented by Kristi Hunziker University of Utah.
Evidence Based Practice RCS /9/05. Definitions  Rosenthal and Donald (1996) defined evidence-based medicine as a process of turning clinical problems.
Learning Objectives In this chapter you will learn about the elements of the research process some basic research designs program evaluation the justification.
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Selecting Evidence Based Practices Oregon’s initial attempts to derive a process Implementation Conversations 11/10.
Quasi-Experimental Designs Slides Prepared by Alison L. O’Malley Passer Chapter 11.
An Expanded Model of Evidence-based Practice in Special Education Randy Keyworth Jack States Ronnie Detrich Wing Institute.
Planning Evaluation Setting the Course Source: Thompson & McClintock (1998) Demonstrating your program’s worth. Atlanta: CDC/National Center for Injury.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Handbook for Health Care Research, Second Edition Chapter 7 © 2010 Jones and Bartlett Publishers, LLC CHAPTER 7 Designing the Experiment.
Replication in Prevention Science Valentine, et al.
Evaluation Research Dr. Guerette. Introduction Evaluation Research – Evaluation Research – The purpose is to evaluate the impact of policies The purpose.
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Open Forum: Scaling Up and Sustaining Interventions Moderator: Carol O'Donnell, NCER
Smoking and other Substance Abuse Prevention Research: History, Standards & Findings Brian R. Flay, D.Phil. Distinguished Professor Public Health and Psychology.
Randomized Control Trials: What, Why, How, When, and Where Mark L. Davison MESI Conference March 11, 2016 Department of Educational Psychology.
Evidence-Based Mental Health PSYC 377. Structure of the Presentation 1. Describe EBP issues 2. Categorize EBP issues 3. Assess the quality of ‘evidence’
Chapter 6 Selecting a Design. Research Design The overall approach to the study that details all the major components describing how the research will.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Clinical Studies Continuum
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Building a Strong Outcome Portfolio
H676 Week 5 - Plan for Today Review your project and coding to date
Presenter: Kate Bell, MA PIP Reviewer
Misc Internal Validity Scenarios External Validity Construct Validity
Presentation transcript:

Standards of Evidence for Prevention Programs Brian R. Flay, D.Phil. Distinguished Professor Public Health and Psychology University of Illinois at Chicago and Chair, Standards Committee Society for Prevention Research Presented at Society for Research on Child Development Atlanta, April

Standards of Evidence Each government agency and academic group that has reviewed programs for lists has its own set of standards. They are all similar but not equal –E.g., CSAP allows more studies in than DE All concern the rigor of the research The Society for Prevention Research recently created standards for the field Our innovation was to consider standards for efficacy, effectiveness and dissemination

Standards for 3 levels: Efficacy, Effectiveness and Dissemination Efficacy –What effects can the intervention have under ideal conditions? Effectiveness –What effects does the intervention have under real-world conditions? Dissemination –Is an effective intervention ready for broad application or distribution?  Desirable –Additional criteria that provide added value to evaluated interventions

Overlapping Standards Efficacy Standards are basic required at all 3 levels Effectiveness Standards include all Efficacy Standards plus others Dissemination standards include all Efficacy and Effectiveness Standards plus others

Four Kinds of Validity (Cook & Campbell, 1979; Shadish, Cook & Campbell, 2002) Construct validity –Program description and measures of outcomes Internal validity –Was the intervention the cause of the change in the outcomes? External validity (Generalizability) –Was the intervention tested on relevant participants and in relevant settings? Statistical validity –Can accurate effect sizes be derived from the study?

Specificity of Efficacy Statement “Program X is efficacious for producing Y outcomes for Z population.” –The program (or policy, treatment, strategy) is named and described –The outcomes for which proven outcomes are claimed are clearly stated –The population to which the claim can be generalized is clearly defined

Program Description Efficacy –Intervention must be described at a level that would allow others to implement or replicate it Effectiveness –Manuals, training and technical support must be available –The intervention should be delivered under the same kinds of conditions as one would expect in the real world –A clear theory of causal mechanisms should be stated –Clear statement of “for whom?” and “under what conditions?” the intervention is expected to work Dissemination –Provider must have the ability to “go-to-scale”

Program Outcomes ALL –Claimed public health or behavioral outcome(s) must be measured Attitudes or intentions cannot substitute for actual behavior –At least one long-term follow-up is required The appropriate interval may vary by type of intervention and state-of-the-field

Measures Efficacy –Psychometrically sound Valid Reliable (internal consistency, test-retest or inter- rater reliability) Data collectors independent of the intervention Effectiveness –Implementation and exposure must be measured Level and Integrity (quality) of implementation Acceptance/compliance/adherence/involvement of target audience in the intervention Dissemination –Monitoring and evaluation tools available

Desirable Standards for Measures For ALL Levels –Multiple measures –Mediating variables (or immediate effects) –Moderating variables –Potential side-effects –Potential iatrogenic (negative) effects

At least one comparison group –No-treatment, usual care, placebo or wait-list Assignment to conditions must maximize causal clarity –Random assignment is “the gold standard” –Other acceptable designs Repeated time-series designs Regression-discontinuity Well-done matched controls –Demonstrated pretest equivalence on multiple measures –Known selection mechanism Design – for Causal Clarity

Level of Randomization In many drug and medical trials, individuals are randomly assigned In educational trials, classrooms or schools must be the unit of assignment –Students within classes/schools are not statistically independent -- they are more alike than students in other classes/schools Need large studies –4 or more schools per condition, preferably 10 or more, in order to have adequate statistical power

Generalizability of Findings Efficacy –Sample is defined Who it is (from what “defined” population) How it was obtained (sampling methods) Effectiveness –Description of real-world target population and sampling methods –Degree of generalizability should be evaluated Desirable Subgroup analyses Dosage studies/analyses Replication with different populations Replication with different program providers

Precision of Outcomes: Statistical Analysis Statistical analysis allows unambiguous causal statements –At same level as randomization and includes all cases assigned to conditions –Test (and adjust fo) for pretest differences –Adjustments for multiple comparisons –Analyses of (and adjustments for) attrition Rates, patterns and types Desirable –Report extent and patterns of missing data

Precision of Outcomes: Statistical Significance Statistically significant effects –Results must be reported for all measured outcomes –Efficacy can be claimed only for constructs with a consistent pattern of statistically significant positive effects –There must be no statistically significant negative (iatrogenic) effects on important outcomes

Precision of Outcomes: Practical Value Efficacy –Demonstrated practical significance in terms of public health (or other relevant) impact –Report of effects for at least one follow-up Effectiveness –Report empirical evidence of practical importance Dissemination –Clear cost information available Desirable –Cost-effectiveness or cost-benefit analyses

Precision of Outcomes: Replication Consistent findings from at least two different high-quality studies/replicates that meet all of the other criteria for efficacy and each of which has adequate statistical power –Flexibility may be required in the application of this standard in some substantive areas When more than 2 studies are available, the preponderance of evidence must be consistent with that from the 2 most rigorous studies Desirable –The more replications the better

Two Reasons for Replication 1. Scientific confidence –All findings must be replicated to avoid chance findings –Preferably by independent evaluators This is very rare so far in Prevention Science 2. Generalizability –Initial efficacy and effectiveness trials usually do not involve representative samples –Need to establish that a program that is effective with one group in one location is effective with others

Additional Desirable Criteria for Dissemination Organizations that choose to adopt a prevention program that barely or not quite meets all criteria should seriously consider undertaking a replication study as part of the adoption effort so as to add to the body of knowledge. A clear statement of the factors that are expected to assure the sustainability of the program once it is implemented.

Embedded Standards Efficacy  Effectiveness  Dissemination  Desirable

Examples of Programs Come Close to Meeting These Standards? Life Skills Training (Botvin) –Multiple RCTs with different populations, implementers and types of training –Only one long-term follow-up –Independent replications of short-term effects are now appearing (as well as some failures) –No independent replications of long-term effects yet Olds’ Home Nursing Program –Multiple replications

Programs for Which the Research Meets the Standards, But Do Not Work DARE –Many quasi-experimental and non-experimental studies suggested effectiveness –Multiple RCTs found no effects (Ennett, et al., 1994 meta-analysis) Hutchinson (Peterson, et al., 2000) –Well-designed RCT –Published results found no long-term effects –But no published information on the program or short-term effects –Published findings cannot be interpreted because of lack of information – they certainly cannot be interpreted to suggest that social influences approaches can never have long-term effects

How the Standards can be used when considering programs Has the program been evaluated in a randomized controlled trial (RCT)? Were classrooms or schools randomized to program and control (no program or alternative program) conditions? Has the program been evaluated on populations like yours? Have the findings been replicated? Were the evaluators independent from the program developers?

Phases of Research in Prevention Program Development (Flay, 1986) 1.Basic Research 2.Hypothesis Development 3.Component Development and Pilot Studies 4.Prototype Studies of Complete Programs 5.Efficacy Trials of Refined Programs 6. Treatment Effectiveness Trials –Generalizability of effects under standardized delivery 7. Implementation Effectiveness Trials –Effectiveness with real-world variations in implementation 8. Demonstration Studies –Implementation and evaluation in multiple systems

School-based Prevention/Promotion Studies are Large and Complex Large randomized trials –With multiple schools or other units per condition Comparisons with “treatment as usual” Measurement of implementation process and program integrity Assessment of effects on presumed mediators –Helps test theories Multiple measures/sources of data –Surveys of students, parents, teachers, staff, community –Teacher and parent reports of behavior –School records for behavior and achievement Multiple, independent trials of promising programs –At both efficacy and effectiveness levels Cost-effectiveness analyses

Much of Prevention is STUCK … We’re Spinning our Wheels At the Efficacy Trial phase –How can we get more programs into effectiveness trials? At the Effectiveness Trial phase –How can we get more proven programs adopted? At the “Model Program” phase –How can we ensure the ongoing effectiveness of model programs? Lots more prevention research is needed – at all levels!