Download presentation
Presentation is loading. Please wait.
Published byStephen Allen Modified over 9 years ago
1
Clemson University School of Ed Faculty Meeting Cindy Van Buren, Ph.D. Division of School Effectiveness August 29, 2014
2
Division of School Effectiveness Office of Educator Services – Mary Hipp Office of Instructional Practices and Evaluations – Briana Timmerman Office of School Leadership – Bruce Moseley Office of School Transformation – Jennifer Morrison Office of Virtual Education – Bradley Mitchell
3
Clemson University School of Ed Faculty Meeting Briana Timmerman, Ph.D. Office of Instructional Practices and Evaluations August 29, 2014
4
Expanded Educator Evaluation System Guidelines http://ed.sc.gov/agency/se/Educator- Evaluation/ http://ed.sc.gov/agency/se/Educator- Evaluation/ Preamble explaining intent and philosophy of teacher professional growth Organized by requirements of ESEA waiver (to prevent restrictions of NCLB)
5
Principal evaluation Current PADEPP Standards 1-9 (50%) School-wide value-add measures (50%)
6
Teacher Evaluation Rubric-based Observations and professional practice ( 50% ) Student growth over school year ( 30% ) Classroom Value-add (tested grades /subjects) Student Learning Objectives (non-tested grades / subjects) District Choice ( 20% )
7
Student Learning Objectives (SLOs) Locally created standards-based student growth target May be school or district-wide or individual for each teacher S pecific, M easureable, A spirational but R ealistic, T ime-specific Uses at least two time periods per student (beginning/end of school year), more data points are better
8
Types of Value-Added Classroom value-added The average growth of all tested students for a given teacher School-wide value-added – OPTIONAL The average growth of all the state-tested students in a school.
9
Value-Add Measures growth rather than achievement 0 = students grew the expected amount Grew more than expected Grew less than expected
10
“Well, that’s fine for general education teachers, but what about special populations?” Honors students vs. struggling students? Students taught by special education teachers?
11
Academic growth is not affected by student abilities
12
Academic growth is not affected by student special needs
13
Your Students All SC students How is that predicted growth calculated?
14
Each individual student’s growth for the year is predicted using the actual growth of other similar students from past years.
15
The 4 th grade scores of students who had 3 rd grade scores that were the same as my student a7 a5 Expected growth is the average growth experienced by similar students. Time Test scores
16
What is Value-added? 2012 achievement2013 achievement References: Meyer & Dokumaci (2009); Wiley (2006) Expected achievement Actual achievement Value-added by that teacher
17
Gain Average Expected Growth – Average Actual Growth So if the gain metric is zero, students met expected growth targets.
18
Default Educator Effectiveness = students met expected growth.
19
2011 2012 Actual = Expected Value-added score = 3 How does this play out in the classroom? Average score of class Mr. Sterling has an average class.
20
2011 2012 Expected Actual Value added score < 3 How does this play out in the classroom? Average score of class Ms. Draper has Honors students.
21
Only 5% of her students scored as “proficient” But 95% made larger gains than expected. Actual Expected Value-added score >3 How does this play out in the classroom? 2011 2012 Ms. Olsen is in a struggling school. Average score of class
22
Less than expected growth, Value-add score is 1or 2 Scores Time VA 5 VA 4 VA 2 VA 1 VA 3 Expected growth = 3 More than expected growth, Value-add score is 4 or 5 Less than expected growth, Value-add score is 1or 2
23
Value-added is associated with positive long- term student outcomes: Improved college attendance—A series of high-value-added teachers may double or even triple college attendance rates. Higher salaries in adulthood—Having one high-value-added teacher is associated with an additional $50K in lifetime earnings per student ($1.5 million for class of 30 students). Reference: Chetty, Friedman, & Rockoff (2011)
24
References Boyd, D., Grossman, P., Lankford, H., Loeb, S., & Wyckoff, J. (2008). Teacher preparation and student achievement (NBER Working Paper Series 14314). Cambridge, MA: National Bureau of Economic Research. Chetty, R., Friedman, J. N., & Rockoff, J. E. (2011). The long-term impacts of teachers: Teacher value-added and student outcomes in adulthood (NBER Working Paper Series 17699). Cambridge, MA: National Bureau of Economic Research. Clotfelter, C. T., Ladd, H. F., & Vigdor, J. L. (2007). Teacher credentials and student achievement in high school: A cross-subject analysis with student fixed effects (NBER Working Paper Series 13617). Cambridge, MA: National Bureau of Economic Research. Goldhaber, D. D., & Brewer, D. J. (1996). Why don’t schools and teachers seem to matter? Assessing the impact of unobservables on educational productivity. Journal of Human Resources, 32 (3), 505–520. Gordon, R., Kane, T., & Staiger, D. O. (2006). Identifying effective teachers using performance on the job (Hamilton Project Discussion Paper). Washington, DC: The Brookings Institution. Harris, D. N., & Sass, T. R. (2009). The effects of NBPTScertified teachers on student achievement. Washington, DC: National Center for Analysis of Longitudinal Data in Education Research. Hershberg, T., Simon, V. A., & Kruger, B. L. (2004). The revelations of value-added. The School Administrator, 61, 10– 14. Koretz, D. (2008). A measured approach. American Educator, Fall, 18–39. Meyer, R. & Dokumaci, E. (2009). Value-added models and the next generation of assessments. Austin, TX: Center for K-12 Assessment & Performance Management. National Association of State Boards of Education (NASBE, 2005). Evaluating value-added: Findings and recommendations from the NASBE Study Group on value-added assessments. Alexandria, VA: Author. Wiley, E. W. (2006). A practitioner’s guide to value-added assessment. Tempe, AZ: Arizona State University.
25
Special considerations for Induction Teachers Encourage student growth to be measured even in first year because it will help the teacher to make better instructional decisions and grow professionally. Teachers in non-tested grades and subjects are likely to need extra assistance with SLOs / data literacy
26
Induction year(s) If a teacher struggles in achieving student growth, recommend additional Induction years. Recommend negative summative evaluation decisions be made with at least three years of growth data.
27
Questions/Discussion
28
Value-Added Measures use Growth not Achievement Why?
29
Growth vs. Achievement Achievement Measures performance at a single point in time. Heavily influenced by family and socio-economic factors. Educators have no control over a student’s incoming achievement status (“uneven playing field”) Growth Compares the same students to themselves over time. Entering achievement level (demographics) don’t affect measure of teacher effectiveness. (“level playing field”)
30
Achievement is affected by demographics
32
Academic growth is not affected by demographics
33
Academic growth is not affected by achievement
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.