Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.

Slides:



Advertisements
Similar presentations
Chapter 22 Evaluating a Research Report Gay, Mills, and Airasian
Advertisements

The Teacher Work Sample
K-6 Science and Technology Consistent teaching – Assessing K-6 Science and Technology © 2006 Curriculum K-12 Directorate, NSW Department of Education and.
Standardized Scales.
Victorian Curriculum and Assessment Authority
Experimental Research Designs
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
8. Evidence-based management Step 3: Critical appraisal of studies
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Magia G. Krause Ph.D. Candidate School of Information University of Michigan Society of American Archivists Annual Meeting August 14, 2009 Undergraduates.
Correlation AND EXPERIMENTAL DESIGN
Who are the participants? Creating a Quality Sample 47:269: Research Methods I Dr. Leonard March 22, 2010.
Impact and outcome evaluation involve measuring the effects of an intervention, investigating the direction and degree of change Impact evaluation assesses.
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Experimental Design The Gold Standard?.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
Student Growth Goals: How Principals can Support Teachers in the Process Jenny Ray PGES Consultant KDE/NKCES.
Determining Sample Size
Collecting Quantitative Data
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.
Day 6: Non-Experimental & Experimental Design
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Measuring Changes in Teachers’ Mathematics Content Knowledge Dr. Amy Germuth Compass Consulting Group, LLC.
Quasi-Experimental Designs For Evaluating MSP Projects: Processes & Some Results Dr. George N. Bratton Project Evaluator in Arkansas.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Classroom Assessments Checklists, Rating Scales, and Rubrics
Slide 1 Long-Term Care (LTC) Collaborative PIP: Medication Review Tuesday, October 29, 2013 Presenter: Christi Melendez, RN, CPHQ Associate Director, PIP.
How to Write a Critical Review of Research Articles
Evaluating a Research Report
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
TWS Aids for Student Teachers & Interns Overview of TWS.
Quantitative and Qualitative Approaches
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
PORTFOLIO ASSESSMENT OVERVIEW Introduction  Alternative and performance-based assessment  Characteristics of performance-based assessment  Portfolio.
Mathematics and Science Partnerships: Summary of the FY2006 Annual Reports U.S. Department of Education.
Primary Purposes of the Evaluation System
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Selecting Evidence Based Practices Oregon’s initial attempts to derive a process Implementation Conversations 11/10.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
What Are the Characteristics of an Effective Portfolio? By Jay Barrett.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Securing External Federal Funding Janice F. Almasi, Ph.D. Carol Lee Robertson Endowed Professor of Literacy University of Kentucky
EBM --- Journal Reading Presenter :呂宥達 Date : 2005/10/27.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Title of your Study Your Name Date of your defense.
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
1 DEMONSTRATION PROJECTS TO ENSURE STUDENTS WITH DISABILITIES RECEIVE A QUALITY HIGHER EDUCATION PROGRAM Performance Measurement, Program and Project Evaluation.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
CONNECT WITH CAEP | | Measures of Teacher Impact on P-12 Students Stevie Chepko, Sr. VP for Accreditation.
Specific Learning Disability: Accurate, Defensible, & Compliant Identification Mississippi Department of Education.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Elayne Colón and Tom Dana
Presenter: Christi Melendez, RN, CPHQ
Understanding Results
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Critical Reading of Clinical Study Results
Performance Improvement Projects: PIP Library
Understanding and Using Standardized Tests
Presenter: Kate Bell, MA PIP Reviewer
Monitoring and Evaluating FGM/C abandonment programs
Presentation transcript:

Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010

2 According to the 2007 Report of the Academic Competitive Council “Successful, large-scale interventions to improve STEM education are unlikely to arise without serious study and trial and error. There is a critical pathway for the development of successful educational interventions and activities, starting generally with small-scale studies to test new ideas and generate hypotheses, leading to increasingly larger and more rigorous studies to test the effect of a given intervention or activity on a variety of students in a variety of settings. Different research methodologies are used along the development pathway, and corresponding evaluation strategies must be used to assess their progress.”

Hierarchy of Study Designs for Evaluating Effectiveness

Overview Criteria for Classifying Designs of MSP Evaluations (“the rubric”) created through the Data Quality Initiative (DQI) Rubric’s key criteria for a rigorous design Common issues with evaluation reports Recommendations for better reporting Discussion of your evaluation challenges and solutions

5 Evaluations Reviewed Using the Rubric All final year evaluations that report using an experimental or quasi-experimental design are considered for review Evaluations need to include a comparison group to ultimately be reviewed with the rubric Within each project, we review evaluations of teacher content knowledge, classroom practices, and student achievement

6 6 criteria used in rubric Rubric comprises six criterion: 1.Equivalence of groups at baseline 2.Adequate sample size 3.Use of valid & reliable measurement instruments 4.Use of consistent data collection methods 5.Sufficient response and retention rates 6.Reporting of relevant statistics

7 Criterion 1 – Baseline Equivalence Requirement Study demonstrates no significant differences in key characteristics between treatment and comparison groups at baseline (for the analytic sample) OR Adequate steps were taken to address the lack of baseline equivalence in the statistical analysis Purpose – Helps rule out alternative explanations for differences between groups

8 Criterion 2 – Sample Size Requirement Sample size is adequate to detect a difference, based on a power analysis using: – Significance level = 0.05, – Power = 0.8 – Minimum detectable effect informed by the literature or otherwise justified Alternatively, meets or exceeds “rule of thumb” sample sizes: – School/district-level interventions: 12 schools – Teacher-level interventions: 60 teachers (teacher outcomes) or 18 teachers (student outcomes) Purpose – Increases the likelihood of finding an impact

9 Criterion 3 – Measurement Instruments Requirement – Data collection instruments used were shown to be valid and reliable to measure key outcomes Use existing instruments that have already been deemed valid and reliable Refer to TCK instrument database developed by MSP Knowledge Management and Dissemination Project at OR Create new instruments that have either been: – Sufficiently tested with subjects comparable to the study sample and found to be valid and reliable, OR – Created using scales and items from pre-existing data collection instruments that have been validated and found to be reliable  Resulting instrument needs to include at least 10 items, and at least 70 percent of the items are from the validated and reliable instrument(s) Purpose – By testing for validity and reliability, you ensure that instruments used accurately capture the intended outcomes

10 Criterion 4 – Data Collection Methods Requirement - Methods, procedures, and timeframes used to collect the key outcome data from treatment and comparison groups are comparable Purpose – Limits possibility that observed differences can be attributed to factors besides the program, such as passage of time and differences in testing conditions

11 Criterion 5 – Attrition Requirement Need to measure key outcomes for at least 70% of original sample (both treatment and control groups), or evidence that attrition is unrelated to treatment If the attrition rates between groups equal or exceed 15 percentage points, difference should be accounted for in the statistical analysis Purpose – Helps ensure that sample attrition does not bias results as participants/control group members who drop out may systematically differ from those who remain

12 Criterion 6 – Relevant Statistics Reported Requirement Include treatment and comparison group post-test means and tests of significance for key outcomes OR, Provide sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Purpose – Provides context for interpreting results, indicating where observed differences between groups are most likely larger than what chance alone might cause

13 Common Issues Found in Evaluation Reports Information critical for complete assessment of all criteria is often not reported, inconsistently reported, or only reported for the treatment group – Pre & post sample sizes for both groups, means, standard deviations/ errors are frequently missing – these are needed for statistical testing and to calculate attrition rates – Varying sample sizes throughout report without explanations for changes – Validity and reliability testing not reported for locally – developed instruments or cited for pre-existing instruments – Data collection methods are not discussed

Key Recommendation – Report the Details Report pre & post sample sizes for both groups and explain changes in samples sizes; if reporting sub-groups, indicate their sample sizes as well Report key characteristics associated with outcomes at baseline (e.g., pretest scores, teaching experience) Document and describe the data collection procedures Report means, standard deviations/errors, for both groups on key outcomes; if using a regression model, describe it Report results from appropriate significance testing of differences observed between groups (e.g., t-statistics or p-values)

Discussion Questions What challenges have you encountered in your efforts to evaluate the MSP project? How have you/might you overcome these obstacles? What has enabled you to increase the rigor of your evaluations? If you could start your evaluation anew, what would you do differently?

Mathematics and Science Partnership (MSP) Programs U.S. Department of Education New Orleans Regional Meeting March 29, 2010