2018 OSEP Project Directors’ Conference

Slides:



Advertisements
Similar presentations
Progress Towards Reading Success: The Reading First Evaluation Prepared by: Amy Kemp, Ph.D. Research Associate and Patricia A. Muller, Ph.D. Associate.
Advertisements

PAYS FOR: Literacy Coach, Power Hour Aides, LTM's, Literacy Trainings, Kindergarten Teacher Training, Materials.
1 Reading First Internal Evaluation Leadership Tuesday 2/3/03 Scott K. Baker Barbara Gunn Pacific Institutes for Research University of Oregon Portland,
Professional Development and Technical Assistance: The National Blueprint for Success Tim Lewis, Ph.D University of Missouri OSEP Center on Positive Behavioral.
First, a little background…  The FIT Program is the lead agency for early intervention services under the Individuals with Disabilities Education Act.
Iowa’s Teacher Quality Program. Intent of the General Assembly To create a student achievement and teacher quality program that acknowledges that outstanding.
FewSomeAll. Multi-Tiered System of Supports A Comprehensive Framework for Implementing the California Common Core State Standards Professional Learning.
“Current systems support current practices, which yield current outcomes. Revised systems are needed to support new practices to generate improved outcomes.”
+ Is your School's Instructional Program Ready for Common Core? Reach Institute for School Leadership.
2015 WI PBIS Conference A5. Benchmark of Quality – Implementation Plan Dave Kunelius - WI RtI Center Regional Coordinator-PBIS.
Collaboration for Effective Educator Development, Accountability and Reform H325A
Data Report July Collect and analyze RtI data Determine effectiveness of RtI in South Dakota in Guide.
Ohio Improvement Process (OIP) Facilitating District-wide Improvement in Instructional Practices and Student Performance.
Technology Action Plan By: Kaitlyn Sassone. What is Systemic Change? "Systemic change is a cyclical process in which the impact of change on all parts.
UPDATE ON EDUCATOR EVALUATIONS IN MICHIGAN Directors and Representatives of Teacher Education Programs April 22, 2016.
Tier 1 Instructional Delivery and Treatment Fidelity Networking Meeting February, 2013 Facilitated/Presented by: The Illinois RtI Network is a State Personnel.
OSEP-Funded TA and Data Centers David Guardino, Office of Special Education Programs, U.S. Department of Education.
Wisconsin Personnel Development System Grant Click on the speaker to listen to each slide. You may wish to follow along in your WPDM Guide.
Module 6: Coaching System
SAM (Self-Assessment of MTSS Implementation) ADMINISTRATION TRAINING
Mgt Project Portfolio Management and the PMO Module 8 - Fundamentals of the Program Management Office Dr. Alan C. Maltz Howe School of Technology.
New Jersey Tiered System of Supports (NJTSS)
TACC Webinar Introduction
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
SOESD’s Teacher Evaluation & Support System
Program Review For School Counseling Programs
Part C Data Managers — Review, Resources, and Relationship Building
Pat Mueller, EEC David Merves, EEC Vitaliy Shyyan, NCEO
Using Formative Assessment
2016 CEEDAR Cross-State Convening
A Workshop for Richland One School District
Refining & Aligning: Recommendations for preparation policy to support rti2 and Special Education in Tennessee Kim Paulsen, vanderbilt university Blake.
Student Services Initiative:
Generate, An Automated federal Reporting Tool
Zelphine Smith-Dixon, State Director of Special Education
9/16/2018 The ACT Government’s commitment to Performance and Accountability – the role of Evaluation Presentation to the Canberra Evaluation Forum Thursday,
RTI & SRBI What Are They and How Can We Use Them?
High-Leverage Practices
Measuring Project Performance: Tips and Tools to Showcase Your Results
DESE Educator Evaluation System for Superintendents
Building a National Collaborative Network for Deaf-Blind Services
2018 OSEP Project Directors’ Conference
2018 OSEP Project Directors’ Conference
2018 OSEP Project Directors’ Conference
Leveraging Evaluation Data: Leading Data-Informed Discussions to Guide SSIP Decisionmaking Welcome Mission of IDC- we provide technical assistance to build.
Implementation Guide for Linking Adults to Opportunity
DISTRICT ACCREDITATION QUALITY ASSURANCE REVIEW
2018 OSEP Project Directors’ Conference
Gifted Education Integrated in RtI Instruction Systems
A3 – Improving State Level Supports and Stakeholder Engagement through Effective Evaluation Kim Gulbrandson, Justyn Poulos – Wisconsin RtI Center Key.
Grantee Guide to Project Performance Measurement
Introductions Introduction
Introduction Introduction
Using Data for Program Improvement
Introduction to Core Professionalism
2018 Improving Data, Improving Outcomes Conference
Implementing Race to the Top
Common Core State Standards AB 250 and the Professional Learning Modules Phil Lafontaine, Director Professional Learning and Support Division.
Parent-Teacher Partnerships for Student Success
Using Data for Program Improvement
Linking Evaluation to Coaching and Mentoring Models
Overview of the Kansas Technical Assistance System Network: Using Technical Assistance to Facilitate Implementation of Evidence-based Practices Kerry.
Georgia’s Tiered System of Supports for Students Karen Suddeth, Project Director Carole Carr, Communications & Visibility Specialist
Bonnie Jortberg, MS,RD,CDE University of Colorado Denver
Part B: Evaluating Implementation Progress and Impact of Professional Development Efforts on Evidence-Based Practices   NCSI Cross-State Learning Collaboratives.
Measuring Child and Family Outcomes Conference August 2008
Access, Equity, and Progress
Using Data to Build LEA Capacity to Improve Outcomes
Want to Integrate Your Data? Let’s Start at the Beginning
Staff Turnover and Silos in Our State, Oh My!
Presentation transcript:

2018 OSEP Project Directors’ Conference OSEP Disclaimer 2018 OSEP Project Directors’ Conference DISCLAIMER: The contents of this presentation were developed by the presenters for the 2018 Project Directors’ Conference. However, these contents do not necessarily represent the policy of the Department of Education, and you should not assume endorsement by the Federal Government. (Authority: 20 U.S.C. 1221e-3 and 3474)

Evaluating Implementation and Improvement: Getting Better Together Moderator: Leslie Fox, OSEP Panelists: Wendy Sawtell, Bill Huennekens, David Merves

Colorado SSIP – Implementation & Improvement Wendy Sawtell State Systemic Improvement Plan Coordinator State Transformation Specialist

ALIGNMENT IS OUR GOAL State-identified Measurable Result is based upon the second portion of the Theory of Action.

Colorado SSIP Plan, Do, Study, Act MODEL FOR IMPROVEMENT What are we trying to accomplish? How will we know that a change is an improvement? What change can we make that will result in improvement? PDSA Cycle and Model for Improvement (1994) Citation: Moen, R. (2009), Foundation and History of the PDSA Cycle, Associates in Process Improvement-Detroit, MI.

6 Core Principles of Improvement Organize as networks (social learning) Learn through disciplined inquiry Embrace measurement See the system Attend to variability Be problem specific and user-centered 6 Core Principles of Improvement Citation: Bryk, A. S. (2018, April 3). Advancing quality in continuous improvement. Speech presented at the Carnegie Foundation Summit on Improvement in Education, San Francisco, CA.

Anthony Bryk, Keynote Speech: Advancing quality in continuous improvement Lesson 1: Take time to really understand the problem you have to solve Lesson 2: Develop evidence that truly informs improvement Lesson 3: As problems increase in their complexity, engage and activate the diversity of expertise assembled in improvement networks Lesson 4: It is a paradigm shift: a cultural transformation in the ways schools work

Practice to Policy Feedback Loop National Implementation Research Network Professional learning, technical assistance, and time for educators to do the work the way it was intended. Changes in practice and policy to produce system change. NIRN - Systemic Change

Fidelity Assessment Measuring Implementation “Fidelity assessment is defined as indicators of doing what is intended.”

SSIP Structured Literacy Project Fidelity Assessment IMPLEMENTATION COACHES Colorado State Systemic Improvement Plan

Phase I, Year One - Appendix J Structured Literacy Routine Rubric Practice Profile Overall Routine Overall understanding of Language Structures Knowledge of Early Reading development Pacing 8 Instructional Components All educators participate in on-going learning opportunities of each element included on the practice profile

Table 1: Method to categorize fidelity assessment items Citation: Retrieved from Module 7 of the Active Implementation Hub on July 6, 2018 Type of Assessment Direct Observation Record Review Ask Others Context Organization of the classroom and student groups Lesson plan is available for review Interview the Master Teacher re: teacher’s planning and preparation activities Content Lesson plan followed during class period Lesson plan contains key elements of an effective approach to teaching the subject Interview Master Teacher re: use of agreed upon curriculum; ratings of Master Teacher regarding reliable inclusion of effective instructional approaches in lesson plans Competence Observation of teachers’ engagement of students with information and questions; and teachers’ use of prompt and accurate feedback Review Coaching notes to see progress on instructional skill development Interview students re: instruction methods used by the teacher

Structured Literacy Routine Rubric Evaluation by Implementation Coaches (3 times per year) CDE Literacy Specialists IMPLEMENTATION COACHES

Comparison in the percentage of students progressing from Kindergarten (2016-2017 SY) to first-grade (2017-2018 SY) in each of the DIBELS Next performance ranges (19 schools)

Table 2: Fidelity scores as a diagnostic tool. Citation: Retrieved from Module 7 of the Active Implementation Hub on July 6, 2018 ? High Fidelity Low Fidelity Good Outcomes Celebrate! Re-examine the innovation and modify the fidelity assessment Poor Outcomes Modify the Intervention Start Over

Our Next Steps Using the 6 Core Principles of Improvement Organize as networks (social learning) Learn through disciplined inquiry Embrace measurement See the system Attend to variability Be problem specific and user-centered 6 Core Principles of Improvement Citation: Bryk, A. S. (2018, April 3). Advancing quality in continuous improvement. Speech presented at the Carnegie Foundation Summit on Improvement in Education, San Francisco, CA.

EXPLORATION PHASE Consideration of a new evaluation tool Network Leadership – Build, Manage, and Evaluate Effective Networks Citation: Retrieved from the Partner Tool website on July 6, 2018

How can I use the Partner tool data? “PARTNER data are meant to be used as a Quality Improvement process, focused on strategic planning (to steer decision-making).” Citation: Retrieved from the Partner Tool website on July 6, 2018

Using the PDSA Cycle with the PARTNER tool “To do this you will need to: identify your goals (plan), implement your collaborative activities (do), gather PARTNER data (study), and develop action steps to get you from where you “are” to where your goals indicate you “should be” (act).” Citation: Retrieved from the Partner Tool website on July 6, 2018

Reflection and Learning To increase your awareness, please read the handout with information about the tool Based upon your knowledge at this point, would this tool be useful to you? Discuss with your elbow partner what may be beneficial and why Or you can access the material directly from the website at https://partnertool.net/about Citation: Retrieved from: the PARTNER Tool website on July 6, 2018

Agenda What is CIID Goal of evaluation efforts Ensure understanding of CIID approach and work, align the plans with that work Execute the plan and collect the data Tools used in the effort Evaluate the data, review recommendations for making changes and improving our work

What is CIID Center for the Integration of IDEA Data Resolve data silos Improve the quality and efficiency or EDFacts reporting Develop and implement automated reporting tool - Generate

Evaluation effort Goal Collaborative effort with external evaluator to establish comprehensive evaluation program to improve Technical Assistance (TA) services to SEAs

Understanding CIID work CIID work is a little different – significant part of the work is IT project management Focus on people, process and systems of data management Evaluation team included in the work – help us shape the work Leadership meetings to calls with TA providers to Training Access to all documentation

Data Collection Plan Establish timeframes Think about work flow in SEAs and with TA efforts Determine methods for different activities Focus groups Manage burden with annual survey Ensure coverage of Universal, Targeted and Intensive TA

Execution of the Plan Collaborative effort Communication – CIID Leadership involved Identify audiences Challenge across special education, IT and EDFacts staff in SEAs

Evaluation Tools Data System Framework Adopted from DaSy Data System Framework used in Part C Administered before and after Determine if goals and objective of data integration were met Intensive Technical Assistance Quality Rubric David will cover in detail

Evaluation Results Review reports in detail with evaluation team Reports include recommendations Share reports with TA providers Explore options for addressing areas that need improvement

Making changes Based on Focus group information – changed engagement with stakeholders Individual meetings with TA providers Trainings with TA providers Example - CEDS Tools

Discussion Question How do we link our collaborative and comprehensive effort to improve services to SEAs to educational outcomes for students with disabilities?

Contact Information Bill Huennekens bill.huennekens@aemcorp.com Follow us on Twitter: @CIIDTA Sign-up for our newsletter: ciidta@aemcorp.com Visit the CIID website: www.ciidta.org Follow us on LinkedIn: https://www.linkedin.com/company/CIID

Thank You!

Evaluating Implementation and Improvement: Getting Better Together Evergreen Evaluation & Consulting, Inc. David Merves MBA, CAS

New Evaluation Tools to Inform Project Improvement Intensive Technical Assistance Quality Rubric PARTNER Tool Measuring Fidelity as a Method for Evaluating Intensive Technical Assistance

Intensive Technical Assistance Quality Rubric Developed by IDEA Data Center (IDC) It is still in DRAFT form Guiding Principles for TA Four Components of Effective Intensive TA

PARTNER TOOL Designed by Danielle Varda with funding from RWJ Foundation Housed at University of Colorado Denver Program to Analyze, Record, and Track Networks to Enhance Relationships Participatory Approach, PARTNER Framework, and FLEXIBLE https://partnertool.net/about/

Measuring Fidelity as a Method for Evaluation Intensive Technical Assistance Presentation at a TACC meeting in September 2015 by Jill Feldman and Jill Lammert of Westat Identify the key components of the intensive TA your Center provides Define and operationalize measurable indicators for key intensive TA components in your logic model Identify data sources and measures for each indicator Create numeric thresholds for determining whether adequate fidelity has been reached for each indicator Assign fidelity ratings (scores) for each indicator and combine indicator scores into a fidelity score for the key component

How would you use any of these tools/instruments in your work? Discussion Question How would you use any of these tools/instruments in your work?

THANK YOU David Merves Evergreen Evaluation & Consulting, Inc david@eecvt.com (802) 857-5935

2018 OSEP Project Directors’ Conference OSEP Disclaimer 2018 OSEP Project Directors’ Conference DISCLAIMER: The contents of this presentation were developed by the presenters for the 2018 Project Directors’ Conference. However, these contents do not necessarily represent the policy of the Department of Education, and you should not assume endorsement by the Federal Government. (Authority: 20 U.S.C. 1221e-3 and 3474)