Encore Webinar February 13, 2019, 3:00pm ET

Slides:



Advertisements
Similar presentations
Researchers as Partners with State Part C and Preschool Special Education Agencies in Collecting Data on Child Outcomes Kathy Hebbeler, ECO at SRI International.
Advertisements

Refresher: Background on Federal and State Requirements.
The Center for IDEA Early Childhood Data Systems ECTA/DaSy System Framework Self- Assessment June 18, 2015.
How to Develop a Project Evaluation Plan Pat Gonzalez Office of Special Education Programs
1 Early Childhood and Accountability OSEP’s Project Director’s Meeting August 2006.
Using State Data to Inform Parent Center Work. Region 2 Parent Technical Assistance Center (PTAC) Conference Charleston, SC June 25, 2015 Presenter: Terry.
Office of Special Education Programs U.S. Department of Education GRANT PERFORMANCE REPORT FOR CONTINUATION FUNDING.
Presented at ECEA-SCASS Meeting Savannah, Georgia October, 2010 OSEP Initiatives on Early Childhood Outcomes Kathy Hebbeler Early Childhood Outcomes Center.
The Center for IDEA Early Childhood Data Systems National Meeting on Early Learning Assessment June 2015 Assessing Young Children with Disabilities Kathy.
Connecticut Part C State Performance Plan Indicator 11 State Systemic Improvement Plan Phase II.
The Center for IDEA Early Childhood Data Systems DaSy Data System Framework: A Tool for Building High-Quality State Part C and Section 619 Data Systems.
Child Outcomes Measurement and Data Quality Abby Winer Schachner & Kathleen Hebbeler International Society on Early Intervention Conference Stockholm,
Dial-in: Passcode: RTI/Multi-Tiered Models of Intervention PLC Movement Between Tiers of Intervention & Implications for Special.
Child Outcomes Summary (COS) Process Professional Development Tools
Child Outcomes Data Collection: Results of the ITCA National Survey and Discussion of Policies and Considerations Christina Kasprzak, ECTA/DaSy Cornelia.
Engaging Families and Creating Trusting Partnerships to Improve Child and Family Outcomes More on Infusing Partnership Principles and Practices into Family.
Child Outcomes Summary Process April 26, 2017
Phase I Strategies to Improve Social-Emotional Outcomes
Part C Data Managers — Review, Resources, and Relationship Building
Pacific and Caribbean States/Entities Early Intervention and
Using Formative Assessment
National, State and Local Educational Environments Data:
Child Outcomes Summary (COS) Process Training Module
OSEP Project Directors Meeting
Supporting Improvement of Local Child Outcomes Measurement Systems
National Webinar Presented by: Amy Nicholas Cathy Smyth
SSIP Evaluation Workshops: Introductory Webinar
2016 Improving Data, Improving Outcomes Conference
Using Data to Reduce Suspension and Expulsion of Young Children
ECTA/DaSy System Framework Self-Assessment
Who Wants to be a 7-Point Rating Scale Millionaire?
ECTA/DaSy System Framework Self-Assessment Comparison Tool
Child Outcomes Data: A Critical Lever for Systems Change
Pay For Success: An Invitation to Learn More
Linking SSIP Implementation and Evaluation: Learning from States
Evaluating Practices Workshop Series
Evaluating Practices Workshop Series
Early Childhood Outcomes Data (Indicator C3 and B7)
2018 OSEP Project Directors’ Conference
IDEA Part C and Part B Section 619 National Child Outcomes Results for
OSEP Initiatives on Early Childhood Outcomes
Assuring the Quality of your COSF Data
Leveraging Evaluation Data: Leading Data-Informed Discussions to Guide SSIP Decisionmaking Welcome Mission of IDC- we provide technical assistance to build.
ECTA/DaSy System Framework Self-Assessment
Cornelia Taylor, Sharon Walsh, and Batya Elbaum
Supporting Improvement of Local Child Outcomes Measurement Systems
2018 Improving Data, Improving Outcomes Conference August 2018
Let’s Talk Data: Making Data Conversations Engaging and Productive
7-Point Rating Scale Jeopardy
Grantee Guide to Project Performance Measurement
Building Capacity to Use Child Outcomes Data to Improve Systems and Practices 2018 DEC Conference.
SSIP Evaluation Workshop 2
Early Childhood and Family Outcomes
OSEP Project Directors Meeting July 2018
NC Preschool Pyramid Model Leadership Team Summit January 9-10, 2019
Researchers as Partners with State Part C and Preschool Special Education Agencies in Collecting Data on Child Outcomes Kathy Hebbeler, ECO at SRI International.
ECO Suggestions on Indicators C3 and B7 Kathy Hebbeler, ECO
Integrating Results into Accountability Procedures and Activities
ECO Suggestions on Indicators C3 and B7 Kathy Hebbeler, ECO
Measuring EC Outcomes DEC Conference Presentation 2010
Child Outcomes Summary (COS) Process Training Module
Refresher: Background on Federal and State Requirements
Part B: Evaluating Implementation Progress and Impact of Professional Development Efforts on Evidence-Based Practices   NCSI Cross-State Learning Collaboratives.
Christina Kasprzak Frank Porter Graham Child Development Institute
Data Culture: What does it look like in your program?
Measuring Child and Family Outcomes Conference August 2008
Early Childhood Outcomes Data (Indicator C3 and B7)
Implementing, Sustaining and Scaling-Up High Quality Inclusive Preschool Policies and Practices: Application for Intensive TA September 10, 2019 Lise.
COS Training to Promote Data Quality: What’s Working, What’s Needed
Data Culture: What does it look like in your program?
Presentation transcript:

Encore Webinar February 13, 2019, 3:00pm ET Evaluating Implementation of Evidence-Based Practices: Tips for Improving Quality and Feasibility Logistics: Attendees are muted upon entry Webinar will be recorded and shared Please type questions into chat or and you will be unmuted Select to expand the chat window You will receive an evaluation Margaret Gillis, DaSy Debbie Shaver, DaSy Katy McCullough, ECTA Stacy Kong, HI Part C Encore Webinar February 13, 2019, 3:00pm ET

Outcomes Participants will increase understanding of Basic evaluation principles related to evaluating practice implementation Features of high-quality tools and methods for evaluating practice implementation Strategies for making evaluation of practice implementation more practical Strategies for summarizing data on practice implementation for different purposes

Agenda What are we measuring and why? How are we measuring? What do we do with all this data? State example: Improving evaluation of Hawaii Part C’s SSIP evidence-based practices

What are we measuring and why?

Practice Implementation Improves Results

Definition of Practices The teachable and doable behaviors that practitioners use with children and families which can be used, replicated, and measured for fidelity The practice needs to be clearly defined in order to be measured.

Definitions Practice Change: Increase or decrease in the number, frequency, precision, or quality of practices a practitioner implements across at least two points in time. Fidelity: The degree to which an intervention or practice is delivered as the developers intended. Fidelity implies strict and continuing faithfulness to the original innovation or practice that is expected to lead to the desired child/family outcomes.

Relationship Between Practice Change and Fidelity Both practice change and practice fidelity are about the implementation of evidence-based practices (EBPs) Are practitioners changing their implementation of EBPs? Are practitioners implementing EBPs as intended? Practitioners can demonstrate changes in practices without reaching fidelity.

Measuring Practice Change and Fidelity Use the same tools or methods to evaluate fidelity versus practice change. Practice change can be evaluated in relation to a fidelity threshold.

Why measure practice change and practice fidelity? To ensure practices are implemented as such that improved outcomes are expected To determine whether practitioners maintain fidelity over time To obtain information for monitoring progress Incremental increases toward fidelity can indicate that improvement strategies are working and highlight areas where practitioners need additional support Corrections/adjustments in practice can be made in a timely manner

Evaluation Plan Components Outcomes Evaluation Questions Performance Indicators Measurement/Data Collection Methods

Alignment of Outcomes, Questions, and Performance Indicators Describe what you intend to achieve as a result of activity(ies) related to EBPs Often interconnected Define steps toward achieving long-term outcomes Evaluation Questions Describe what you need to know to determine if you have achieved the outcome Performance Indicators Describe how you will determine if you have met the outcome Based on measurement data (e.g., extant or new)

Example: Are these aligned? Outcome Description Evaluation Question Performance Indicator Providers use evidence-based practices to support social-emotional skills with parents and infants/toddlers identified with social emotional needs Do providers know how to use evidence-based practices to support social-emotional skills with parents and infants/toddlers identified with social-emotional needs? 80% of providers completed training on evidence-based practices to support social emotional skills with parents and infants/toddlers identified with social-emotional needs  

Example: Alignment Outcome Description Evaluation Question Performance Indicator Providers use evidence-based practices to support social-emotional skills with parents and infants/toddlers identified with social emotional needs Do providers use evidence-based practices to support social-emotional skills with parents and infants/toddlers identified with social-emotional needs? 80% of providers use evidence-based practices to support social emotional skills with parents and infants/toddlers identified with social-emotional needs

How are we measuring?

Defining and Operationalizing Evidence-based Practices Practices that are not clearly defined and operationalized are not measureable A Practice Profile is a way to define and articulate acceptable and non acceptable practices (National Implementation Research Network) A Practice Profile is not a measurement tool but can be used as a starting place to develop or identify a tool for assessing practice change and fidelity

Balancing High-Quality and Practical Measurement The higher the quality of the measurement approach, the more useful and actionable the data are. The key is to ensure you get the most meaningful, accurate, and reliable information possible. Measurement also needs to be doable, practical

Characteristics of High-Quality Measurement The tool: Is Aligned with the selected evidence-based practices Provides Valid information– accurate information on practice implementation Is Reliable – produces consistent information across users, settings, activities, and time points Is Practical – can be used with the staff and resources available

Characteristics of High-Quality Measurement The tool: Provides a meaningful fidelity threshold score that indicates whether practitioners have reached a level of implementation that is sufficient for achieving targeted child or family outcomes. Captures variation (e.g., over time, across practitioners) Provides useful information to practitioners to identify areas of strength and areas for improvement to move toward, reach, and maintain fidelity.

Characteristics of High-Quality Measurement Other considerations to improve quality: Timing/frequency. The tool is used at sufficient frequency to measure incremental improvement (practice change) and the maintenance of fidelity over time. Training. Clear instructions and training are provided to improve data quality. Purpose of data collection is clear - Practitioners and data collectors understand how data will be used to improve implementation.  

Data Collection Methods Observation by someone else Practitioner self report Review of documents Multiple methods can be used, including qualitative methods

How to Make it Practical Leverage existing processes and structures (e.g., supervisors, coaches) Start with existing tools (e.g., fidelity tools associated with a model, tools related to DEC Recommended Practices) Prioritize the practices to be evaluated—it is not practical to evaluate all practices Start small—select a subgroup of practitioners and/or programs for more intensive measures (e.g., observations)

What do we do with all this data?

Purposes of Data Analysis and Use What do you want to know from your data? What information do decision-makers at each system level need to have to improve practice implementation? The type of decisions made at each level drives the evaluation questions and the appropriate analysis.

Example Questions at Different System Levels Practitioner Level Is the practitioner implementing with fidelity? Local Program Level Are practitioners implementing the EBP with fidelity? Which components of the fidelity tool show the greatest growth across practitioners? The least growth? State Level Are practitioners within local programs implementing the EBP with fidelity? What percentage of programs meet a performance indicator for percentage of practitioners at fidelity?

Data Aggregation Example

Hawaii’s SiMR Infants and toddlers with disabilities, in demonstration sites, will have substantially increased their rate of growth in social-emotional skills (including social relationships) by the time they exit early intervention. Summary Statement 1: Of those children who entered or exited the program below age expectations, the percent who substantially increased their rate of growth by the time they turned 3 years of age or exited the program.

Hawaii’s SSIP Broad Improvement Strategies Professional development (PD) around Primary Service Provider (PSP) Approach to Teaming and Coaching model to support social-emotional (SE) development with fidelity. Increase the capacity of EI programs to provide services and supports to address SE development. Enhance the child outcomes summary (COS) process

Hawaii’s EBPs & Measurement Tools Primary Service Provider Approach to Teaming *Checklist for Implementing a Primary – Coach Approach to Teaming (measures infrastructure) Coaching Model SE Self-Assessment (measures practice change) *Coaching Log Summary Form (measures fidelity) *Hawaii SE Competencies Coaching Log Review (in process) COS Process Monitoring Tool *Shelden & Rush

Shifts to Better Align Evaluation Added a short-term outcome that was previously missed but vital to the work being done. Deleted a short-term outcome that was embedded into another outcome. Deleted an intermediate outcome because it was time intensive to track and not necessary.  For all evaluation outcomes elaborated on who will enter data, how will data be stored, how will data be entered, when data will be entered, how data will be transmitted included analysis descriptions

Hawaii’s Evaluation Plan - example Outcome Evaluation Question Performance Indicator Measurement EI providers will report improved quality implementation of EBP (coaching model in natural learning environments) to support SE development. Do Demonstration Site Providers report improvements in the quality of implementation of EBP (coaching model in natural learning environments) to support SE development? 75% of Demonstration Site Providers will report being at Level III: Triadic Relationships for at least 7 competencies on the SE Competency Self-Assessment. Tool: SE Competency Self-Assessment (SECSA) tracking sheet Data collection: CSPD coordinator completes the SECSA tracking sheet with ratings for all demo site providers Intervals: Before training, 6 mo. after training, 12 mo. after training, then annually

Questions/Comments?

Key Take-Away Points Alignment, alignment, alignment Practices = behaviors High-quality measurement = meaningful, useful data Independent observation = more objective information than practitioners’ self report Needs to be practical Summarize data for different systems levels

Resources DaSy and ECTA tip sheets on evaluating practice implementation forthcoming Resources and Tools section of ECTA Statewide Implementation Guide http://ectacenter.org/sig/ Strengthening SSIP Evaluations with Qualitative Methods https://dasycenter.org/strengthening-ssip-evaluations-with-qualitative-methods/

Contact Information Margaret.gillis@sri.com Katy.mccullough@unc.edu Debbie.shaver@sri.com Stacy.Kong@doh.Hawaii.gov

Thank you! Visit the DaSy website at: http://dasycenter.org/ Follow DaSy on Twitter: @DaSyCenter Visit the ECTA website at: http://ectacenter.org/ Follow ECTA on Twitter: @ECTACenter

Thank you The contents of this tool and guidance were developed under grants from the U.S. Department of Education, #H326P120002 and #H326P170001. However, those contents do not necessarily represent the policy of the U.S. Department of Education, and you should not assume endorsement by the Federal Government. Project Officers: Meredith Miceli, Richelle Davis, and Julia Martin Eile.