The hybrid success model: Theory and practice G. Gage Kingsbury Martha S. McCall Northwest Evaluation Association A paper presented to the Seminar on longitudinal.

Slides:



Advertisements
Similar presentations
Testing for Tomorrow Growth Model Testing Measuring student progress over time.
Advertisements

Mark D. Reckase Michigan State University The Evaluation of Teachers and Schools Using the Educator Response Function (ERF)
Iowa Assessment Update School Administrators of Iowa November 2013 Catherine Welch Iowa Testing Programs.
Standardized Tests What They Measure How They Measure.
Implementing Virginia’s Growth Measure: A Practical Perspective Deborah L. Jonas, Ph.D. Executive Director, Research and Strategic Planning Virginia Department.
Fall 2014 MAP NWEA Score Comparison Alliance Dr. Olga Mohan High School October 22, 2014.
Measures of Academic Progress (MAP) Curt Nath Director of Curriculum Ocean City School District.
Evaluating Pretest to Posttest Score Differences in CAP Science and Social Studies Assessments: How Much Growth is Enough? February 2014 Dale Whittington,
Enquiring mines wanna no.... Who is it? Coleman Report “[S]chools bring little influence to bear upon a child’s achievement that is independent of.
Robert L. Linn CRESST, University of Colorado at Boulder Paper presented at a symposium sponsored by the National Association of Test Directors entitled.
Classroom Assessment A Practical Guide for Educators by Craig A
1 The New York State Education Department New York State’s Student Reporting and Accountability System.
Vertical Scale Scores.
Reporting college and career readiness results to the public DQC Public Reporting Task Force | January 9, 2014.
Using Data to Improve Student Achievement Summer 2006 Preschool CSDC.
NCLB AND VALUE-ADDED APPROACHES ECS State Leader Forum on Educational Accountability June 4, 2004 Stanley Rabinowitz, Ph.D. WestEd
John Cronin, Ph.D. Director The Kingsbury NWEA Measuring and Modeling Growth in a High Stakes Environment.
Measuring Charter Quality Eric Paisner, NAPCS Anna Nicotera, NAPCS Lyria Boast, Public Impact.
Introduction to the Georgia Student Growth Model Understanding and Using SGPs to Improve Student Performance 1.
SB : The Great Teachers and Leaders Act State-wide definition of “effective” teacher and principal in Colorado Academic growth, using multiple measures.
Benchmark Data. World History Average Score: 56% Alliance: 96%
Classroom Assessments Checklists, Rating Scales, and Rubrics
PPT Presentation Template: This PPT includes all slides to present a district or building level overview of PVAAS. This was used with a district- wide.
The Impact of Including Predictors and Using Various Hierarchical Linear Models on Evaluating School Effectiveness in Mathematics Nicole Traxel & Cindy.
Review and Validation of ISAT Performance Levels for 2006 and Beyond MetriTech, Inc. Champaign, IL MetriTech, Inc. Champaign, IL.
Teacher Evaluation System Part II: Student Learning Data May 5, 2015.
Instruction, Teacher Evaluation and Value-Added Student Learning Minneapolis Public Schools November,
1 Watertown Public Schools Assessment Reports 2010 Ann Koufman-Frederick and Administrative Council School Committee Meetings Oct, Nov, Dec, 2010 Part.
An Introduction to Student-Involved Assessment FOR learning, Rick Stiggins Chapter 11: Report Cards Assessments of Learning.
Final Reports from the Measures of Effective Teaching Project Tom Kane Harvard University Steve Cantrell, Bill & Melinda Gates Foundation.
Growth Model for District “X” Why Use Growth Models? Showing progress over time is a more fair way of evaluating It is not just a “snap shot” in time.
Issues in Selecting Assessments for Measuring Outcomes for Young Children Issues in Selecting Assessments for Measuring Outcomes for Young Children Dale.
© 2007 Board of Regents of the University of Wisconsin System, on behalf of the WIDA Consortium WIDA Focus on Growth H Gary Cook, Ph.D. WIDA.
Guide to Test Interpretation Using DC CAS Score Reports to Guide Decisions and Planning District of Columbia Office of the State Superintendent of Education.
Standard Setting Results for the Oklahoma Alternate Assessment Program Dr. Michael Clark Research Scientist Psychometric & Research Services Pearson State.
Pearson Copyright 2010 Some Perspectives on CAT for K-12 Assessments Denny Way, Ph.D. Presented at the 2010 National Conference on Student Assessment June.
Scaling and Equating Joe Willhoft Assistant Superintendent of Assessment and Student Information Yoonsun Lee Director of Assessment and Psychometrics Office.
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Annual Measurable Objectives (trajectory targets).
Mathematics and Science Partnerships Program Improving Math and Science Achievement in Low-Performing, High-Poverty Schools: Implications for Professional.
DVAS Training Find out how Battelle for Kids can help Presentation Outcomes Learn rationale for value-added progress measures Receive conceptual.
PVAAS Overview: Evaluating Growth, Projecting Performance PVAAS Statewide Core Team Fall 2008.
1 Getting Up to Speed on Value-Added - An Accountability Perspective Presentation by the Ohio Department of Education.
Measures of Academic Progress (MAP) Presenter: Rachel Iaconelli-Scheyer Director of Curriculum Director of Curriculum Ocean City School District Ocean.
The Normal Distribution and Norm-Referenced Testing Norm-referenced tests compare students with their age or grade peers. Scores on these tests are compared.
Public School Accountability System. Uses multiple indicators for broad picture of overall performance Uses multiple indicators for broad picture of overall.
LaKenji Hastings, NWLC Assessment Program Specialist Georgia Milestones Parent Informational.
Balancing on Three Legs: The Tension Between Aligning to Standards, Predicting High-Stakes Outcomes, and Being Sensitive to Growth Julie Alonzo, Joe Nese,
Curriculum Night Elementary. What do I as a parent need to know to support student assessments at CCAS? Essential Question.
KHS PARCC/SCIENCE RESULTS Using the results to improve achievement Families can use the results to engage their child in conversations about.
Curriculum Night Middle School. What do I as a parent need to know to support student assessments at CCAS? Essential Question.
To support efforts to raise student achievement To support the district’s accountability status To offer standardized accountability metrics to complement.
Using Data to Improve Student Achievement Summer 2006 Preschool CSDC.
USING MAP DATA TO SET GOALS AND ENGAGE STUDENTS AND FAMILIES 7/14/16.
MAP Growth NWEA Northwest Evaluation Association.
Nuts and Bolts of Assessment
Growth: Changing the Conversation
Measures of Academic Progress (MAP) – Overview
IT’S ALL ABOUT GROWTH!. Hemet Unified School District’s Use of Measures of Academic Progress (MAP)
NWEA Measures of Academic Progress (MAP)
Booklet Design and Equating
2015 PARCC Results for R.I: Work to do, focus on teaching and learning
Data-Based Instructional Decision Making
School Performance Measure Calculations SY
Measures of Academic Progress (MAP)
Formative Assessments Director, Assessment and Accountability
Assessment Literacy: Test Purpose and Use
Office of Strategy, Innovation and Performance
Virginia Board of Education’s
Curriculum and Assessment
Presentation transcript:

The hybrid success model: Theory and practice G. Gage Kingsbury Martha S. McCall Northwest Evaluation Association A paper presented to the Seminar on longitudinal measurement, University of Maryland April, 2005

What does “success” mean for a school? Consider two schools, both of which have half of the students in each grade identified as proficient or higher Which one is more successful? Consider that one of these schools caused twice as much growth as the other Which one is more successful?

Difficulties with Proficiency Models A model which counts the number of students in a few proficiency categories causes schools to focus on borderline students, to the possible detriment of all other students The setting of proficiency levels is an exercise which can cause inconsistent and inappropriate decisions

Difficulties with Growth Models Growth models need strong assessments to operate well If students grow, it doesn’t mean they will become proficient

A Hybrid Success Model Set a growth target for each student that will lead to proficiency Measure each student’s status Measure each student’s growth Judge a school by the success of each student

Elements of a Hybrid Success Model Observed Growth = X t2 – X t1 Expected Growth = the average growth of students with a beginning achievement level equal to this student. Standards Growth = the amount of growth needed to get to proficiency by a target year. (Proficiency ty – X t1 )/ (number of yrs to target) Growth Target = the greater of Expected Growth and Standards Growth

Hybrid Success Index (HSI) for a student HSI = 1 if Observed Growth >= Growth Target Otherwise HSI = Observed Growth / Growth Target

Hybrid Success Value: School or District HSV = Mean HSI

Advantages of the HSM Incorporates both proficiency levels and growth Gives credit for growth of all students Sets predetermined goals for both students and schools Gives a more complete picture of school and district growth

Comparison of HSM and Traditional Value- Added Models HSM Adds substantial information to standards approach Identifies schools that cause more growth than typical Growth of every student counts Compares schools to a constant scale Requires a single measurement scale All schools can be successful Points out needed instruction Needed growth to succeed is known in advance Value-Added Model Adds substantial information to standards approach Identifies schools that cause more growth than typical Growth of every student counts Compares schools to one another Allows multiple measures Always results in high and low schools Doesn’t point out needed instruction Growth isn’t prespecified

What is Needed to Implement the Hybrid Success Model? Accurate assessments A measurement scale that allows growth across years A vertical scale related to proficiency levels across grades A process to assign useful growth targets to students

What is needed to implement a vertical measurement scale that can measure growth across years? Very accurate measurement for each student A data system that follows students across time At least one, preferably two or more scores per year. A theory of learning that allows continuous growth A scale that doesn’t change across time

Vertical scaling is enhanced by adaptive testing Each form is unique Each score has small error Item parameter estimates have less error Links throughout the scale are dense and accurate

 Vertical Linking Block  Grade X Form Grade X +1 Form Fixed Form Vertical Linking

 Grade X  Grade X +1 Adaptive Continuous Vertical Linking

What is meant by an equal interval scale? For any 2 values of theta* on the scale, the odds ratio of success on a given item equals the odds ratio of the two scores. A 1-unit change in theta difference results in a unit change in the odds for success These scales do NOT indicate growth measures that are equal for equal intervals of time.

When conditions for vertical scaling are not met, results can be disappointing Trying to merge two or more existing scales is not advised (e.g., merging existing benchmark scales). Merging scales from tests given far apart in time can be difficult to interpret (e.g. Haertel’s analysis of NAEP scales) Fixed form linking is often too weak for vertical scaling (e.g., Huynh, Meyer & Barton)

Advantages of Using Scale-related Growth Measures Scores can be interpreted in terms of curriculum continuum descriptors Provides practical knowledge of what the student has mastered and what needs to come next Growth can be shown both absolutely and in comparison to performance standards Meaning of growth is more accessible to teachers, parents and students

Growth is not uniform at all ages or abilities

Growth is not uniform in reading either

A Comparison of Three Models --149,000 students --74,788 8 th graders in ,302 5 th graders in Reading & mathematics scores from Spring 2003, Spring 2004, Spring 2005

Models of accountability compared --Percent meeting standard median of most recent state studies (smoothed across grade levels) --Standardized growth ((Observed growth - expected Growth) / SD of Exp Growth) using point growth norms --Hybrid success value Using a proportionate growth increment toward 10 th grade standards

Accountability for Schools MeasureSchool CriteriaPolicy Goal Percent Meeting Standard>= 57% % needed to meet 100% proficiency in 2014 Standardized Growth>=0 Greater than expected so that mean moves upward on the scale Hybrid Success>=.771 Index needed to meet 100% proficiency in 2014

Rate of School Success Schools/Grades Meeting Accountability Criteria AllGr 5 MathGr 8 MathGr 5 ReadGr 8 Read ModelN%N%N%N%N% % Met173464%57663%27762%56463%31770% GrowZ131749%41846%18642%42047%29365% HSV180567%64771%25156%61768%29064%

Agreement between Model Accountability Criteria Agreement AllGr 5 MathGr 8 MathGr 5 ReadGr 8 Read %Met/GrowZ/HSV N%N%N%N%N% Y/Y/Y101037%31535%13630%32736%23251% Y/Y/N181%00%41%20%123% Y/N/Y53320%21824%9120%19121%337% Y/N/N1736%435%4610%445%409% N/Y/Y1847%748%215%677%225% N/Y/N1054%293%256%243%276% N/N/Y783%404%31%324%31% N/N/N61323%19421%12227%21424%8318% Total %913100%448100%901100%452100%

Final thoughts Students, teachers, and administrators need accountability goals that they can see beforehand and tie to specific action Accountability that doesn’t require growth for all students is weak accountability Accountability that doesn’t reward students and teachers for great growth is weak accountability

Thank You for Your Thoughtful Consideration of These Ideas Northwest Evaluation Association