Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008.

Slides:



Advertisements
Similar presentations
Educational Research: Causal-Comparative Studies
Advertisements

Chapter 22 Evaluating a Research Report Gay, Mills, and Airasian
Defining Characteristics
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Experimental Research Designs
Quasi-Experimental Design
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Educational Research by John W. Creswell. Copyright © 2002 by Pearson Education. All rights reserved. Slide 1 Chapter 11 Experimental and Quasi-experimental.
Validity 4 For research to be considered valid, it must be based on fact or evidence; that is, it must be "capable of being justified."
Quantitative Research  Adequate experimental control  Lack of artificiality  Basis for comparison  Adequate information from the data  Uncontaminated.
Experimental Research
Experimental Research
Nasih Jaber Ali Scientific and disciplined inquiry is an orderly process, involving: problem Recognition and identification of a topic to.
Evaluation of Math-Science Partnership Projects (or how to find out if you’re really getting your money’s worth)
Experimental Design The Gold Standard?.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
I want to test a wound treatment or educational program in my clinical setting with patient groups that are convenient or that already exist, How do I.
Research Design for Quantitative Studies
Day 6: Non-Experimental & Experimental Design
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Measuring Changes in Teachers’ Mathematics Content Knowledge Dr. Amy Germuth Compass Consulting Group, LLC.
Methodology Describe Context & setting Design Participants Sampling Power Analysi s Interventions Outcome (study variables) Data Collection Procedures.
NERLYNE ST. PIERRE WILBERTA WILLIAM EDUCATION 7202T – SPRING 2011 A Combination of Teacher-Centered and Student-Centered Classroom Management Approaches.
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
Evaluating a Research Report
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Assisting GPRA Report for MSP Xiaodong Zhang, Westat MSP Regional Conference Miami, January 7-9, 2008.
Chapter Four Experimental & Quasi-experimental Designs.
Types of Research and Designs This week and next week… Covering –Research classifications –Variables –Steps in Experimental Research –Validity –Research.
Quantitative and Qualitative Approaches
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar April 25, 2012 Note: These slides are intended as.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
MSP Annual Performance Report: Online Instrument MSP Regional Conferences November, 2006 – February, 2007.
Classifying Designs of MSP Evaluations Lessons Learned and Recommendations Barbara E. Lovitts June 11, 2008.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Chapter 3 should describe what will be done to answer the research question(s), describe how it will be done and justify the research design, and explain.
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
November 15, Regional Educational Laboratory - Southwest The Effects of Teacher Professional Development on Student Achievement: Finding from a Systematic.
Mathematics and Science Partnerships: Summary of the FY2006 Annual Reports U.S. Department of Education.
Primary Purposes of the Evaluation System
Experimental Research
The Evaluation of Mathematics and Science Partnerships Program A Quasi Experimental Design Study Abdallah Bendada, Title II Director
Chapter 10 Common Experimental Research Designs
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Research Design ED 592A Fall Research Concepts 1. Quantitative vs. Qualitative & Mixed Methods 2. Sampling 3. Instrumentation 4. Validity and Reliability.
U.S. Department of Education Mathematics and Science Program State Coordinators’ Meeting.
1 Module 3 Designs. 2 Family Health Project: Exercise Review Discuss the Family Health Case and these questions. Consider how gender issues influence.
Evaluation Designs Adrienne DiTommaso, MPA, CNCS Office of Research and Evaluation.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Securing External Federal Funding Janice F. Almasi, Ph.D. Carol Lee Robertson Endowed Professor of Literacy University of Kentucky
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Chapter Eight: Quantitative Methods
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Effectiveness of Selected Supplemental Reading Comprehension Interventions: Impacts on a First Cohort of Fifth-Grade Students June 8, 2009 IES Annual Research.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
June 25, Regional Educational Laboratory - Southwest Review of Evidence on the Effects of Teacher Professional Development on Student Achievement:
Research designs Research designs Quantitative Research Designs.
Chapter 12 Quantitative Questions and Procedures.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Chapter Eight: Quantitative Methods
Designing an Experiment
Grants Academy Session Three
Quantitative Research
CCSSO National Conference on Student Assessment June 21, 2010
Reminder for next week CUELT Conference.
Presentation transcript:

Overview of MSP Evaluation Rubric Gary Silverstein, Westat MSP Regional Conference San Francisco, February 13-15, 2008

Data Quality Initiative Technical assistance to ED grant programs Help individual ED programs refine GPRA goals Help individual ED programs refine GPRA goals Develop procedures and tools to help projects document progress in meeting GPRA goals Develop procedures and tools to help projects document progress in meeting GPRA goals Currently working with 25 ED programs Currently working with 25 ED programs DQI tools developed For MSP MSPTCK software (for GPRA measure on teacher content knowledge MSPTCK software (for GPRA measure on teacher content knowledge Evaluation rubric (for GPRA measure on study design) Evaluation rubric (for GPRA measure on study design)

MSP GPRA Measures for Evaluation Design The percentage of MSP projects that use an experimental or quasi-experimental design for their evaluations The percentage of MSP projects that use an experimental or quasi-experimental design for their evaluations that are conducted successfully and that yield scientifically valid results

Study Designs that Meet the GPRA Goal Experimental study—the study measures the intervention’s effect by randomly assigning individuals (or other units, such as classrooms or schools) to a group that participated in the intervention, or to a control group that did not, and then compares post-intervention outcomes for the two groups Quasi-experimental study—the study measures the intervention’s effect by comparing post-intervention outcomes for treatment participants with outcomes for a comparison group (that was not exposed to the intervention), chosen through methods other than random assignment Comparison-group study with equating—a study in which statistical controls and/or matching techniques are used to make the treatment and comparison groups similar in their pre-intervention characteristics Comparison-group study with equating—a study in which statistical controls and/or matching techniques are used to make the treatment and comparison groups similar in their pre-intervention characteristics Regression-discontinuity study—a study in which individuals (or other units, such as classrooms or schools) are assigned to treatment or comparison groups on the basis of a “cutoff” score on a pre-intervention non-dichotomous measure Regression-discontinuity study—a study in which individuals (or other units, such as classrooms or schools) are assigned to treatment or comparison groups on the basis of a “cutoff” score on a pre-intervention non-dichotomous measure

MSP Evaluation Rubric Tool for reviewing MSP evaluations that used an experimental or quasi-experimental design Rubric delineates the minimum criteria for an MSP evaluation to have (1) been successfully conducted and (2) yielded valid data An evaluation has to meet each criterion in order to meet the GPRA measure The criteria can also be used by projects and evaluators to assure that the documentation of their study design is aligned with the GPRA goal

Rubric Components For Experimental Designs: Sample size Sample size Quality of measurement instruments Quality of measurement instruments Quality of data collection methods Quality of data collection methods Data reduction rates—i.e., attrition and response Data reduction rates—i.e., attrition and response Relevant statistics reported Relevant statistics reported For Quasi-experimental Designs: All of the above, plus All of the above, plus Baseline equivalence of groups Baseline equivalence of groups

Component #1: Sample Size Met the criterion—sample size for the overall study was adequate Based on power analysis with recommended significance level=0.05, power=0.8, and a minimum detectable effect informed by the literature or otherwise justified Based on power analysis with recommended significance level=0.05, power=0.8, and a minimum detectable effect informed by the literature or otherwise justified Did not meet the criterion—the sample size was too small Did not address the criterion

Component #2: Quality of Measurement Instruments Met the criterion The study used existing data collection instruments that had already been deemed valid and reliable to measure key outcomes, OR The study used existing data collection instruments that had already been deemed valid and reliable to measure key outcomes, OR Data collection instruments developed specifically for the study were sufficiently pre-tested with subjects who were comparable to the study sample Data collection instruments developed specifically for the study were sufficiently pre-tested with subjects who were comparable to the study sample Did not meet the criterion The key data collection instruments used in the evaluation lacked evidence of validity and reliability The key data collection instruments used in the evaluation lacked evidence of validity and reliability Did not address the criterion

Component #3: Quality of Data Collection Methods Met the criterion The methods, procedures, and timeframes used to collect key outcome data from treatment and control groups were the same The methods, procedures, and timeframes used to collect key outcome data from treatment and control groups were the same Did not meet the criterion Instruments/assessments were administered differently in manner to treatment and control group participants, OR Instruments/assessments were administered differently in manner to treatment and control group participants, OR Instruments/assessments were administered at different times to treatment and control group participants Instruments/assessments were administered at different times to treatment and control group participants Did not address the criterion

Component #4: Data Reduction Rates—Met the Criterion The study measured the key outcome variable(s) in the post-tests for at least 70% of the original study sample (treatment and control groups combined) OR there is evidence that the high rates of data reduction were unrelated to the intervention AND The proportion of the original study sample that was retained in follow-up data collection activities (e.g., post-intervention surveys) and/or for whom post-intervention data were provided (e.g., test scores) was similar for both the treatment and control groups (i.e., less or equal to a 15-percent difference), OR the proportion of the original study sample that was retained in the follow-up data collection was different for the treatment and control groups, but sufficient steps were taken to address this differential attrition in the statistical analysis

Component #4: Data Reduction Rates—Did Not Meet Criterion The study failed to measure the key outcome variable(s) in the post- tests for 30% or more of the original study sample (treatment and control groups combined), and there is no evidence that the high rates of data reduction were unrelated to the intervention OR The proportion of study participants who participated in follow-up data collection activities (e.g., post-intervention surveys) and/or for whom post-intervention data were provided (e.g., test scores) was significantly different for the treatment and control groups (i.e., more than a 15-percent difference) and sufficient steps to address differential attrition were not taken in the statistical analysis

Component #5: Relevant Statistics Reported Met the criterion The final report includes treatment and control group post-test means and tests of statistical significance for key outcomes or provides sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) The final report includes treatment and control group post-test means and tests of statistical significance for key outcomes or provides sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Did not meet the criterion The final report does not include treatment and control group post-test means and/or tests of statistical significance for key outcomes or provide sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) The final report does not include treatment and control group post-test means and/or tests of statistical significance for key outcomes or provide sufficient information for calculation of statistical significance (e.g., mean, sample size, standard deviation/standard error) Did not address the criterion

Component #6: Baseline Equivalence of Groups (for QED) Met the criterion There were no significant pre-intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, or adequate steps were taken to address the lack of baseline equivalence in the statistical analysis There were no significant pre-intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, or adequate steps were taken to address the lack of baseline equivalence in the statistical analysis Did not meet the criterion There were statistically significant pre-intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, and no steps were taken to address lack of baseline equivalence in the statistical analysis There were statistically significant pre-intervention differences between treatment and comparison group participants on variables related to the study’s key outcomes, and no steps were taken to address lack of baseline equivalence in the statistical analysis Did not address the criterion

Questions? Breakout Session on GPRA Reporting Thursday, February 14 from 10:15 am to 11:45 am Sienna II—3 rd floor