Holli G. Bayonas, Ph.D & Eric S. Howard, M.A.

Slides:



Advertisements
Similar presentations
P-20 Data Collaborative Grant University/College Work Group February 24, 2010.
Advertisements

Mywish K. Maredia Michigan State University
TEACHER QUALITY AND DISTRIBUTION Principals and Teachers Effectiveness and Evaluation NSBA’s Federal Relations Network Conference February
Explaining Race Differences in Student Behavior: The Relative Contribution of Student, Peer, and School Characteristics Clara G. Muschkin* and Audrey N.
Teacher Quality, Distribution, and Turnover in El Paso Ed Fuller The University of Texas at Austin El Paso, Tx June28, 2006.
The World Bank Human Development Network Spanish Impact Evaluation Fund.
Districts and States Working with VARC Minneapolis Milwaukee Racine Chicago Madison Tulsa Atlanta New York City Los Angeles Hillsborough County NORTH DAKOTA.
S-STEM Program Evaluation S-STEM PI Meeting Arlington, VA October 2012.
Drawing by Mankoff: copyright 1993 The New Yorker Magazine, Inc.
Working Toward a Statewide Information System to Track the Effectiveness of Student Aid Financial Programs in Maryland Michael J. Keller Director of Policy.
Teacher-Designed Incentive Pay in Texas A Presentation to the IES Research Conference by Lori L. Taylor.
Evaluating NSF Programs
Vouchers in Milwaukee: What Have We Learned From the Nation’s Oldest and Largest Program? Deven Carlson University of Oklahoma.
Leveraging Race to the Top to Maximize the Use of Data To Ensure College & Career Readiness Aimee R. Guidera Achieve ADP September 10, 2009.
Evaluating the Vermont Mathematics Initiative (VMI) in a Value Added Context H. ‘Bud’ Meyers, Ph.D. College of Education and Social Services University.
C.O.R.E Creating Opportunities that Result in Excellence.
The Impact of Including Predictors and Using Various Hierarchical Linear Models on Evaluating School Effectiveness in Mathematics Nicole Traxel & Cindy.
Quasi Experimental Methods I Nethra Palaniswamy Development Strategy and Governance International Food Policy Research Institute.
Impacts of Comprehensive Teacher Induction: Final Results from a Randomized Trial IES Summer Research Conference, June 2010 Steven Glazerman ● Eric Isenberg.
Mathematics and Science Education U.S. Department of Education.
CONSORTIUM FOR EDUCATIONAL RESEARCH AND EVALUATION – NORTH CAROLINA The NC Race to the Top Evaluation Plan: An Introduction October 10, 2011 Gary T. Henry,
How Much Value is Added? An Evaluation Plan for the Achievement Challenge Pilot Project.
Need For The Project Jennifer Doolittle, Ph.D. April 23, 2009.
AFRICA IMPACT EVALUATION INITIATIVE, AFTRL Africa Program for Education Impact Evaluation David Evans Impact Evaluation Cluster, AFTRL Slides by Paul J.
ESEA, TAP, and Charter handouts-- 3 per page with notes and cover of one page.
Randomized Assignment Difference-in-Differences
Teacher Incentive Fund U.S. Department of Education.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Measuring Turnaround Success October 29 th, 2015 Jeanette P. Cornier, Ph.D.
Do European Social Fund labour market interventions work? Counterfactual evidence from the Czech Republic. Vladimir Kváča, Czech Ministry of Labour and.
Patricia Gonzalez, OSEP June 14, The purpose of annual performance reporting is to demonstrate that IDEA funds are being used to improve or benefit.
Revamping the Teaching Profession by Attracting Non-Teachers to It: Evidence from Enseña Chile Mariana Alfonso Education Division, Inter-American Development.
Research Opportunities in AMSP UK Mathematics Education Retreat October 15, 2005.
Classroom Network Technology as a Support for Systemic Mathematics Reform: Examining the Effects of Texas Instruments’ MathForward Program on Student Achievement.
Higher Education Act.
Looking for statistical twins
Student Affairs: A Culture of Assessment
Rockledge High School: Turnaround Story
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Improving Educator Practice through Instructional Mentoring
Merit & Incentive Pay Based on High Stakes Testing
Equitable Access to Excellent Educators
L. Elia, A. Morescalchi, G. Santangelo
Quasi Experimental Methods I
Experimental Design-Chapter 8
Excellent Educators for All Initiative
School Quality and the Black-White Achievement Gap
Evidence of a Program's Effectiveness in Improving Colorectal Cancer Screening Rates in Federally Qualified Health Centers Robert L. Stephens, PhD, MPH1;
New Teachers Induction/Retention EDU 7527 Mr. Tim Martin
CCSD Mentoring & Induction
The University of Texas System Assessment of
Defining and Measuring Student Success Dr
FY17 Evaluation Overview: Student Performance Rating
Dr. Robert H. Meyer Research Professor and Director
Impact evaluation: The quantitative methods with applications
Matching Methods & Propensity Scores
Examining the Effects of New Teacher Induction
Propensity Score Matching Makes Program Evaluation Easy
Matching Methods & Propensity Scores
Module 7: Monitoring Data: Performance Indicators
Using Data to Improve Student Achievement Aimee R. Guidera
Matching Methods & Propensity Scores
North Carolina Positive Behavior Support Initiative
Impact Evaluation Methods: Difference in difference & Matching
Evaluating Impacts: An Overview of Quantitative Methods
Student Equity Planning August 28, rd Meeting
Tell A Meaningful Story With Data Through Research
David Mann David Stapleton (Mathematica Policy Research) Alice Porter
Measuring Student Growth
State Examples and Follow-up Data Requests for SOQ Proposals
Presentation transcript:

Holli G. Bayonas, Ph.D. & Eric S. Howard, M.A. The SERVE Center at the University of North Carolina at Greensboro Propensity Score Matching for the Evaluation of a Federal Teacher Incentive Fund Grant Program Theory Background The theory of change for the TIF-funded program is that financial incentives and professional development will result in a larger pool of highly qualified applicants, higher retention rates of teachers, and better school climate. Long-term outcomes include increased student achievement. Evaluation Questions: 1. To what extent are the most highly qualified teachers and administrators being recruited and retained at TIF Schools? 2. To what extent were the TIF teachers and administrators trained as proposed? 3. To what extent did the training and incentives impact teacher and student outcomes? The data used in this evaluation includes climate survey results, teacher retention data, teacher data such as certification and years of experience, and various student data such as graduation rates, discipline data, and achievement data. The propensity score matching was implemented as a way to determine if growth in student achievement was a result of the program or a result of other unexplained factors. Propensity scores were calculated for schools within North Carolina but outside of the district using the following variables: 2005-06 Enrollment numbers, 2005-06 1-year teacher turnover rate, 2005-06 % free and reduced lunch, 2005-06 % minority, 2005-06 # of teachers with 0-3 years of experience, 2004-05 Performance Composite, and 2005-06 Performance Composite. Essentially, these were the same variable that were examined by administrators when determining which schools would become the TIF schools. A nearest neighbor approach was used when selecting 3 comparison schools to every 1 treatment school. Presenting Treatment and Comparison School Results Background PSM PSM Advantages: (1) it allows the researcher to identify non-comparable covariates early in the analysis; (2) it is less sensitive to model misspecification; and (3) the research is allowed to generate a non-parsimonious model (Hahs-Vaughn & Onwuegbuzie, 2006). PSM Disadvantages: (1) the model only considers and controls for observed data; (2) differences between treatment and control groups based on their unobservable characteristics are not controlled for; (3) the need for larger data sets to get optimal performance; and (4) the inclusion of irrelevant covariates will reduce model efficiency (Hahs-Vaughn & Onwuegbuzie, 2006). Abstract The SERVE Center at UNCG was awarded a 5-year contract to evaluate 30 schools receiving a federal Teacher Incentive Fund (TIF) grant. Select teachers in hard to fill position are given a recruitment incentive and also a performance incentive when their students make gains in achievement (measured by SAS Value-Added). Using propensity score matching (PSM) and publicly available data, comparison schools were selected across North Carolina. In the Year 2-5 evaluation reports, these comparison schools will be essential to see if trends witnessed in the TIF schools are treatment effects or if these trends are shared by the comparison schools. Presenting variables for the treatment school along with its three comparison schools in chart form facilitates evidence-based decision-making by stakeholders. Summary 1. Without comparison groups, stakeholders could incorrectly conclude results of a program. While stakeholders are concerned about closing the gap on other schools within the county, they need to be able to accurately attribute effects on students. When scores go up in a treatment school and in some comparison schools, then the program cannot take credit for increased student achievement. When there is no comparison school, and variables of interest move in a positive direction, stakeholders can be tempted to attribute this to the program and not take into account other possible influences. 2. When we aggregate for t-tests and other statistics, we loose valuable individual school data which shows the program working better in some schools than in others. Presenting all of the variables in line charts along with the comparison schools allows a more in-depth look at the school, which is essential for decision-making. 3. Identifying control groups through Propensity Score Matching has methodological limitations but it should be considered as an addition to reporting trend data solely on the treatment schools. Performance Composite Example: In which of these three schools can we say that the program is contributing positively to increasing overall achievement? Should we be looking at only one variable when determining effectiveness? Only one variable was statistically significant after the matches; 1-year teacher turnover rate. The remaining variables were considered equal across treatment and control groups. The resulting sample consisted of 28 treatment schools and 70 comparison schools. For the main program report, t-tests were calculated to determine differences between control and treatment. For decision-making in Year 2, individual school reports were created so that stakeholders could see a school’s performance in relation to three schools that did not receive the treatment.