Examining Value-Added Models to Measure Teacher Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.

Slides:



Advertisements
Similar presentations
Value Added in CPS. What is value added? A measure of the contribution of schooling to student performance Uses statistical techniques to isolate the.
Advertisements

Testing for Tomorrow Growth Model Testing Measuring student progress over time.
Teacher Evaluation and Rewards OECD Mexico Joint Workshop December 1-2, 2009 Susan Sclafani National Center on Education and the Economy.
On The Road to College and Career Readiness Hamilton County ESC Instructional Services Center Christina Sherman, Consultant.
VALUE – ADDED 101 Ken Bernacki and Denise Brewster.
Implementing Virginia’s Growth Measure: A Practical Perspective Deborah L. Jonas, Ph.D. Executive Director, Research and Strategic Planning Virginia Department.
MEASURING TEACHING PRACTICE Tony Milanowski & Allan Odden SMHC District Reform Network March 2009.
Understanding and Addressing Achievement Gaps 2014 Title 1 Directors Conference This work was originally produced in whole or in part by American Institutes.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Teacher Evaluation Models: A National Perspective Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination,The National.
How can we measure teachers’ contributions to student learning growth in the “non-tested” subjects and grades? Laura Goe, Ph.D. Research Scientist, ETS,
C R E S S T / U C L A Improving the Validity of Measures by Focusing on Learning Eva L. Baker CRESST National Conference: Research Goes to School Los Angeles,
Using Student Learning Growth as a Measure of Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
+ Hybrid Roles in Your School If not now, then when?
Teacher Evaluation Systems: Opportunities and Challenges An Overview of State Trends Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
John Cronin, Ph.D. Director The Kingsbury NWEA Measuring and Modeling Growth in a High Stakes Environment.
Educator Preparation, Retention, and Effectiveness Ed Fuller University Council for Educational Administration and The University of Texas at Austin February.
PUBLIC SCHOOLS OF NORTH CAROLINA STATE BOARD OF EDUCATION DEPARTMENT OF PUBLIC INSTRUCTION 1 Review of the ABCs Standards SBE Issues Session March 2, 2005.
Measures of Teachers’ Contributions to Student Learning Growth Laura Goe, Ph.D. Research Scientist, Understanding Teaching Quality Research Group, ETS.
1 National Council of Supervisors of Mathematics Illustrating the Standards for Mathematical Practice: Getting Started with the Practices Investigations.
National Center on Educational Outcomes Fall, 2004 Alternate assessment, gaps, and other challenges! A view of current practices from the technical assistance.
Measuring Teacher Effectiveness: Challenges and Opportunities
Evaluating the Vermont Mathematics Initiative (VMI) in a Value Added Context H. ‘Bud’ Meyers, Ph.D. College of Education and Social Services University.
Jim Lloyd_2007 Educational Value Added Assessment System (EVAAS) Olmsted Falls City Schools Initial Presentation of 4 th Grade Students.
1 New York State Growth Model for Educator Evaluation 2011–12 July 2012 PRESENTATION as of 7/9/12.
Update on Virginia’s Growth Measure Deborah L. Jonas, Ph.D. Executive Director for Research and Strategic Planning Virginia Department of Education July-August.
RESEARCH ON MEASURING TEACHING EFFECTIVENESS Roxanne Stansbury EDU 250.
Educator Evaluation Spring Convening Connecting Policy, Practice and Practitioners May 28-29, 2014 Marlborough, Massachusetts.
Measuring teacher effectiveness using multiple measures Ellen Sullivan Assistant in Educational Services, Research and Education Services, NYSUT AFT Workshop.
What Are We Measuring? The Use of Formative and Summative Assessments Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and.
Reaching Every Student with an Excellent Teacher Presentation to Project L.I.F.T. October 7, 2011.
Elaine Weiss, National Coordinator, BBA AASA Legislative Advocacy Conference Washington, DC July 12, 2011 ESEA: Getting Accountability and Assessment Right.
Evaluating Teacher Effectiveness: Selecting Measures Laura Goe, Ph.D. SIG Schools Webinar August 12, 2011.
What Works? What Doesn’t? Overview of Teacher Compensation: What Works? What Doesn’t? James H. Stronge College of William and Mary Williamsburg, Virginia.
“Value added” measures of teacher quality: use and policy validity Sean P. Corcoran New York University NYU Abu Dhabi Conference January 22, 2009.
TEACHER EFFECTIVENESS INITIATIVE VALUE-ADDED TRAINING Value-Added Research Center (VARC)
1 New York State Growth Model for Educator Evaluation 2011–12 July 2012 PRESENTATION as of 7/9/12.
Models for Evaluating Teacher Effectiveness Laura Goe, Ph.D. California Labor Management Conference May 5, 2011  Los Angeles, CA.
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
DVAS Training Find out how Battelle for Kids can help Presentation Outcomes Learn rationale for value-added progress measures Receive conceptual.
Copyright © 2009 National Comprehensive Center for Teacher Quality. All rights reserved. Student Learning and Achievement in Measuring Teacher Effectiveness.
State Practices for Ensuring Meaningful ELL Participation in State Content Assessments Charlene Rivera and Lynn Shafer Willner GW-CEEE National Conference.
CALIFORNIA STATE BOARD OF EDUCATION State Policies: Orchestrating the Common Core Mathematics Classroom Ilene W. Straus, Vice President California State.
Race to the Top Assessment Program: Public Hearing on Common Assessments January 20, 2010 Washington, DC Presenter: Lauress L. Wise, HumRRO Aab-sad-nov08item09.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to National Conference of State Legislatures July 14, 2011  Webinar.
ANNOOR ISLAMIC SCHOOL AdvancEd Survey PURPOSE AND DIRECTION.
Release of Preliminary Value-Added Data Webinar August 13, 2012 Florida Department of Education.
Using EVAAS to Improve Student Performance Donna Albaugh Rachel McBroom Heather Stewart Region 4 PD Leads NCDPI.
EVAAS and Expectations. Answers the question of how effective a schooling experience is for learners Produces reports that –Predict student success –Show.
Teacher Evaluation Systems 2.0: What Have We Learned? EdWeek Webinar March 14, 2013 Laura Goe, Ph.D. Research Scientist, ETS Sr. Research and Technical.
Teacher Quality/Effectiveness: Defining, Developing, and Assessing Policies and Practices Part III: Setting Policies around Teacher Quality/Effectiveness.
Weighting components of teacher evaluation models Laura Goe, Ph.D. Research Scientist, ETS Principal Investigator for Research and Dissemination, The National.
C R E S S T / CU University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Design Principles for Assessment.
Wisconsin Administrative Code PI 34 1 Wisconsin Department of Public Instruction - Elizabeth Burmaster, State Superintendent Support from a Professional.
C R E S S T / CU University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Measuring Adequate Yearly.
Measures of Teachers’ Contributions to Student Learning Growth Laura Goe, Ph.D. Research Scientist, Understanding Teaching Quality Research Group, ETS.
Forum on Evaluating Educator Effectiveness: Critical Considerations for Including Students with Disabilities Lynn Holdheide Vanderbilt University, National.
Student Growth Model Salt Lake City School District Christine Marriott Assessment and Evaluation Department Salt Lake City School District State.
Making Data Work for Kids: EVAAS Teacher Reports October 2012 SAS ® EVAAS ® for K-12.
 Mark D. Reckase.  Student achievement is a result of the interaction of the student and the educational environment including each teacher.  Teachers.
Using Student Growth in Teacher Evaluation and Development Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive.
1 New York State Growth Model for Educator Evaluation June 2012 PRESENTATION as of 6/14/12.
Evaluating Teacher Effectiveness Laura Goe, Ph.D. Presentation to the Hawaii Department of Education July 20, 2011  Honolulu, HI.
Evaluating Teacher Effectiveness: Where Do We Go From Here?
Evaluating Teacher Effectiveness: An Overview Laura Goe, Ph.D.
Teacher Evaluation Models: A National Perspective
2015 PARCC Results for R.I: Work to do, focus on teaching and learning
Dr. Robert H. Meyer Research Professor and Director
Federal Policy & Statewide Assessments for Students with Disabilities
Teacher Evaluation: The Non-tested Subjects and Grades
Presentation transcript:

Examining Value-Added Models to Measure Teacher Effectiveness Laura Goe, Ph.D. Research Scientist, ETS, and Principal Investigator for the National Comprehensive Center for Teacher Quality Hofstra University Doctoral Policy Forum October 15, 2011  Hempstead, NY 1

2 The goal of teacher evaluation The ultimate goal of all teacher evaluation should be… TO IMPROVE TEACHING AND LEARNING

3 Trends in teacher evaluation Policy is way ahead of the research in teacher evaluation measures and models  Though we don’t yet know which model and combination of measures will identify effective teachers, many states and districts are compelled to move forward at a rapid pace Inclusion of student achievement growth data represents a huge “culture shift” in evaluation  Communication and teacher/administrator participation and buy-in are crucial to ensure change The implementation challenges are enormous  Few models exist for states and districts to adopt or adapt  Many districts have limited capacity to implement comprehensive systems, and states have limited resources to help them

4 How did we get here? Value-added research shows that teachers vary greatly in their contributions to student achievement (Rivkin, Hanushek, & Kain, 2005). The Widget Effect report (Weisberg et al., 2009) “…examines our pervasive and longstanding failure to recognize and respond to variations in the effectiveness of our teachers.” (from Executive Summary)

5 A concise definition of teacher effectiveness Anderson (1991) stated that “… an effective teacher is one who quite consistently achieves goals which either directly or indirectly focus on the learning of their students” (p. 18).

6 Validity and use of assessments to evaluate teachers Tests, systems, etc. do not have validity Validity lies in how they are used  A test designed to measure student knowledge and skills in a specific grade and subject may be valid for determining where that student is relative to his/her peers at a given point in time  However, there are questions about validity in terms of using such test results to measure teachers - What part of a student’s score is attributable solely to the teacher’s instruction and effort?

7 Growth vs. Proficiency Models End of YearStart of School Year Achievement Proficient Teacher B: “Failure” on Ach. Levels Teacher A: “Success” on Ach. Levels In terms of growth, Teachers A and B are performing equally Slide courtesy of Doug Harris, Ph.D, University of Wisconsin-Madison

8 Growth vs. Proficiency Models (2) End of YearStart of School Year Achievement Proficient Teacher A Teacher B A teacher with low- proficiency students can still be high in terms of GROWTH (and vice versa) Slide courtesy of Doug Harris, Ph.D, University of Wisconsin-Madison

9 Most popular growth models: Value-added and Colorado Growth Model EVAAS uses prior test scores to predict the next score for a student Teachers’ value-added is the difference between actual and predicted scores for a set of students mlhttp:// ml Colorado Growth model  Betebenner 2008: Focus on “growth to proficiency”  Measures students against “academic peers” 

10 Slide courtesy of Damian Betebenner at Linking student learning results to professional growth opportunities

11 What value-added and growth models cannot tell you Value-added and growth models are really measuring classroom, not teacher, effects Value-added models can’t tell you why a particular teacher’s students are scoring higher than expected  Maybe the teacher is focusing instruction narrowly on test content  Or maybe the teacher is offering a rich, engaging curriculum that fosters deep student learning. How the teacher is achieving results matters!

12 Value-Added: Student effects “A teacher who teaches less advantaged students in a given course or year typically receives lower-effectiveness ratings than the same teacher teaching more advantaged students in a different course or year.” “Models that fail to take student demographics into account further disadvantage teachers serving large numbers of low-income, limited English proficient, or lower-tracked students.” (Newton et al., 2010, pg 2)

13 Value-Added: Error rates and stability “Type I and II error rates for comparing a teacher’s performance to the average are likely to be about 25 percent with three years of data and 35 percent with one year of data.” “Any practical application of value-added measures should make use of confidence intervals in order to avoid false precision, and should include multiple years of value-added data in combination with other sources of information to increase reliability and validity.” (Schochet & Chiang, 2010, abstract)

14 Value-Added: Subscales Teachers’ scores on subscales of a test can yield very different results, which also raises the question of weighting subscale results (Lockwood et al, 2007)  Lockwood et al. found substantial variation in teachers’ rankings based on the subscales (“Problem Solving” and “Procedures”)  More variation within teachers than across teachers “Our results provide a clear example that caution is needed when interpreting estimated teacher effects because there is the potential for teacher performance to depend on the skills that are measured by the achievement tests” (Lockwood et al, 2007, pg. 55)

15 Value-Added: Test content Polikoff and colleagues (2011) found that  About half of standards are tested - If half the standards they are teaching are not tested, how can the test accurately reflect teachers’ contribution to student learning?  About half of test content corresponds with grade/subject standards - If half of test content is material that is not in the standards teachers are supposed to be teaching, is if fair to hold teachers accountable for test results?

16 Value-Added: Multiple teachers In one study, 21% of teachers in Washington, DC had students who had also been in another math teacher’s class that year (Hock & Isenberg, 2011)  This covered all situations, including students who had changed classes or schools as well as co-teaching and other cases where students were taught by more than one teacher  Hock & Isenberg determined best estimates were obtained by counting student multiple times, for each teacher the student had, rather than trying to account for how much each teacher contributed to students’ scores

17 Value-Added: Possible responses to technical challenges Use multiple years of data to mitigate sorting bias and gain stability in estimates (Koedel & Betts, 2009; McCaffrey et al., 2009; Glazerman et al., 2010 ) Use confidence intervals and other sources of information to improve reliability and validity of teacher effectiveness ratings (Glazerman et al., 2010) Have teachers and administrators verify rosters to ensure scores are calculated with students the teachers actually taught Consider the importance of subscores in teacher rankings

18 Growth Models Wisconsin’s Value-Added Research Center (VARC) SAS Education Value-Added Assessment System (EVAAS) Mathematica mpr.com/education/value_added.asp American Institutes of Research (AIR) Colorado Growth Model

19 References Betebenner, D. W. (2008). A primer on student growth percentiles. Dover, NH: National Center for the Improvement of Educational Assessment (NCIEA). Braun, H., Chudowsky, N., & Koenig, J. A. (2010). Getting value out of value-added: Report of a workshop. Washington, DC: National Academies Press. Glazerman, S., Goldhaber, D., Loeb, S., Raudenbush, S., Staiger, D. O., & Whitehurst, G. J. (2011). Passing muster: Evaluating evaluation systems. Washington, DC: Brown Center on Education Policy at Brookings. Glazerman, S., Goldhaber, D., Loeb, S., Raudenbush, S., Staiger, D. O., & Whitehurst, G. J. (2010). Evaluating teachers: The important role of value-added. Washington, DC: Brown Center on Education Policy at Brookings. Herman, J. L., Heritage, M., & Goldschmidt, P. (2011). Developing and selecting measures of student growth for use in teacher evaluation. Los Angeles, CA: University of California, National Center for Research on Evaluation, Standards, and Student Testing (CRESST).

20 References (continued) Hock, H., & Isenberg, E. (2011). Methods for accounting for co-teaching in value-added models. Princeton, NJ: Mathematica Policy Research. Teaching%20in%20VAMs.pdf Koedel, C., & Betts, J. R. (2009). Does student sorting invalidate value-added models of teacher effectiveness? An extended analysis of the Rothstein critique. Cambridge, MA: National Bureau of Economic Research. McCaffrey, D., Sass, T. R., Lockwood, J. R., & Mihaly, K. (2009). The intertemporal stability of teacher effect estimates. Education Finance and Policy, 4(4), Linn, R., Bond, L., Darling-Hammond, L., Harris, D., Hess, F., & Shulman, L. (2011). Student learning, student achievement: How do teachers measure up? Arlington, VA: National Board for Professional Teaching Standards. Lockwood, J. R., McCaffrey, D. F., Hamilton, L. S., Stecher, B. M., Le, V.-N., & Martinez, J. F. (2007). The sensitivity of value-added teacher effect estimates to different mathematics achievement measures. Journal of Educational Measurement, 44(1),

21 References (continued) New York State Education Department (2011). Summary of Provisions in 3012c Regulations: May, 2011 (revised September 14, 2011 for impact of August Court decision and other clarifications) Newton, X. A., Darling-Hammond, L., Haertel, E., & Thomas, E. (2010). Value-added modeling of teacher effectiveness: An exploration of stability across models and contexts. Education Policy Analysis Archives, 18(23). Polikoff, M. S. (2011). How well aligned are state assessments of student achievement with state content standards? American Educational Research Journal, 48(4), Policy Analysis for California Education and Rennie Center for Education Research and Policy (2011). The road ahead for state assessments. Cambridge, MA: Rennie Center for Education Research and Policy. Race to the Top Application Rivkin, S. G., Hanushek, E. A., & Kain, J. F. (2005). Teachers, schools, and academic achievement. Econometrica, 73(2),

22 References (continued) Sanders, W. L., & Horn, S. P. (1998). Research findings from the Tennessee Value-Added Assessment System (TVAAS) Database: Implications for educational evaluation and research. Journal of Personnel Evaluation in Education, 12(3), Schochet, P. Z., & Chiang, H. S. (2010). Error rates in measuring teacher and school performance based on student test score gains. Washington, DC: National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education. Weisberg, D., Sexton, S., Mulhern, J., & Keeling, D. (2009). The widget effect: Our national failure to acknowledge and act on differences in teacher effectiveness. Brooklyn, NY: The New Teacher Project.

23 Questions?

24 Laura Goe, Ph.D. P: Website: