To Consortia, or not to Consortia

Slides:



Advertisements
Similar presentations
ADVANCES IN TECHNOLOGY-BASED TRAINING Bradford S. Bell and Steve W. J. Kozlowski.
Advertisements

1 Gifted and Talented Task Force Presentation to the Howell Township Board of Education March 29, 2006.
PARCC Tests: An Investment in Learning Test quality and rigor increase; Costs for states generally hold steady July 2013.
Common Core Standards and the Edmonds School District November 4, 2013.
On The Road to College and Career Readiness Hamilton County ESC Instructional Services Center Christina Sherman, Consultant.
1 Designing High Quality, Affordable Assessment Systems Edward Roeber Michigan State University National Research Council Board on Testing and Assessment.
CCSS-M IN MICHIGAN. Michigan K-12 Standards CCSS-M with Michigan “Welcome”
What Are States Doing to Prepare For the Next Generation of Assessments? Planning For and Beyond John Olson Barry Topol National Conference on.
KRISTEN BURTON ERIN FAASUAMALIE Future of Alternate Achievement Standards and Assessment in Wisconsin Wisconsin Department of Public Instruction.
Introduction to the Smarter Balanced Digital Library Brian Huff September 3, 2014.
Optimum Solutions Corporation All the Right Answers.
Common Core State Standards & Assessment Update The Next Step in Preparing Michigan’s Students for Career and College MERA Spring Conference May 17, 2011.
Welcome What’s a pilot?. What’s the purpose of the pilot? Support teachers and administrators with the new evaluation system as we learn together about.
The Five New Multi-State Assessment Systems Under Development April 1, 2012 These illustrations have been approved by the leadership of each Consortium.
SMARTER Balanced Assessment Consortium Smarter Balanced Assessment Consortium.
Consortia of States Assessment Systems Instructional Leaders Roundtable November 18, 2010.
Ohio’s Assessment Future The Common Core & Its Impact on Student Assessment Evidence by Jim Lloyd Source doc: The Common Core and the Future of Student.
PARCC Update June 6, PARCC Update Today’s Presentation:  PARCC Field Test  Lessons Learned from the Field Test  PARCC Resources 2.
Salem-Keizer Public Schools Budget Message.
TRANSPORTATION MANAGEMENT
High Quality Assessments – Supporting state goals and policies while developing a balanced assessment system Angela Hemingway Director, Assessment and.
 Participants will teach Mathematics II or are responsible for the delivery of Mathematics II instruction  Participants attended Days 1, 2, and 3 of.
ASG Cost Model Presentation CCSSO National Conference on Student Assessment.
Direct Math Assessment State Department Training September, 2008 Cindy Johnstone Mathematics Coordinator.
Getting the Most Value for Your Assessment Dollar – Designing Adapting and Maintaining Quality Assessment Programs During Tough Economic Times To Consortia,
1 Assessment of the Common Core State Standards CCCOE Curriculum Council January, 2011.
Smarter Balanced Assessment System March 11, 2013.
May 13, 2011 Getting to Know the Common Core State Standards (CCSS)
Getting the Most Value for Your Assessment Dollar: Cost Analysis for the New Kentucky Assessment System John F. Olson Assessment Solutions Group CCSSO.
Amity School of Business ADVANCES IN TECHNOLOGY-BASED TRAINING.
Planning for What States are Doing to Prepare for the Next Generation of Assessments Barry Topol John Olson National Conference on Student Assessment.
Pearson Copyright 2010 Some Perspectives on CAT for K-12 Assessments Denny Way, Ph.D. Presented at the 2010 National Conference on Student Assessment June.
Next Generation Assessments Stakeholder Meeting June 26, 2014.
ASG Cost Model Presentation New Jersey Department of Education October 21, 2009.
PARENT UNIVERSITY: Preparing Your Child for Success Presented by: Dr. Alicia Scelso, Principal, Pequannock Township High School Richard M. Hayzler, Principal,
Standardized Testing EDUC 307. Standardized test a test in which all the questions, format, instructions, scoring, and reporting of scores are the same.
LaKenji Hastings, NWLC Assessment Program Specialist Georgia Milestones Parent Informational.
SBCUSD and Smarter Balanced Assessment Consortium Testing Assessment and Accountability, January 2013.
KHS PARCC/SCIENCE RESULTS Using the results to improve achievement Families can use the results to engage their child in conversations about.
Designing High Quality Assessments that are Affordable: Conclusions and Recommendations John F. Olson CCSSO National Conference on Student Assessment Detroit,
Understanding Standards: Advanced Higher Statistics
Strategic Information Systems Planning
Writing Objectives Understand the ways in which the writing assessment differs from other assessments Administer the Writing Performance Assessment.
Measuring Impact Guide
ADVANCES IN TECHNOLOGY-BASED TRAINING
WEBINAR The Total Economic Impact Of Software-As-A-Service
American Evaluation Association Conference November 10, 2010
Analyzing Performance Tasks: Turning Results Into Action
MANAGING HUMAN RESOURCES
Smarter Balanced Assessment Results
Cypress Ridge Elementary School
Shared Services Open Forum
2015 PARCC Results for R.I: Work to do, focus on teaching and learning
Considerations of Content Alignment in CAT
Connecticut Core Standards for Mathematics
Student Mobility and Achievement Growth In State Assessment Mohamed Dirir Connecticut Department of Education Paper presented at National Conference.
Assessments aligned to Common Core State Standards
California Assessment of Student Performance and Progress and the California Alternate Assessments Special Education Local Plan Area Directors May 5,
Kathy Cox State Superintendent of Schools GPS Day 3 Training
Creating Student Learning Objectives (SLOs)
Quality Education Commission Presentation to the Senate Committee
Gateway High School-Alt.Ed Annual Title 1 Parent Workshop
Chapter 14 Sourcing Decisions in a Supply Chain
New Assessments and Accommodations
Deciding the mixed-mode design WP1
Getting the Most Value for Your Assessment Dollar: Cost Analysis for the New Kentucky Assessment System John F. Olson Assessment Solutions Group CCSSO.
The Assessment Consortia An Overview of the Designs of PARCC
Presentation transcript:

To Consortia, or not to Consortia Getting the Most Value for Your Assessment Dollar – Designing Adapting and Maintaining Quality Assessment Programs During Tough Economic Times To Consortia, or not to Consortia CCSSO National Conference on Student Assessment June 23, 2010

Joining a Consortium Implementing a new, innovative assessment program in a consortium as a way to save costs. Or, maintaining a current program without having to make drastic cuts Is it doable? Can a consortia of states implement a new assessment at a significantly reduced cost than a single state acting alone? How large does the consortia need to be? Where are the cost savings opportunities? June 23, 2010 www.assessmentgroup.org

Joining a State Assessment Consortium Joining a state assessment consortium can have its advantages but. . . Requires a lot of planning, coordination & desire Several successful examples: NECAP WIDA Achieve Algebra 2 PARCC & SBAC (responses due today) June 23, 2010

Stanford/Nellie Mae Study Purpose of the study was to see if it is possible to create an affordable “high quality” assessment Step one – Model a current typical assessment in ELA and Math – Cost $19-$20 Step two – Model a high quality assessment for the same state – Cost $55-$56 a student Step three – Implement several cost savings strategies June 23, 2010 Assessment Solutions Group www.assessmentgroup.org

Cost Reduction Strategies Participation in a consortium Looked at 10, 20 and 30 state sizes Cost reduction - $15 per student Uses of technology for online test delivery, distributed human scoring of some of the open-ended items, and automated scoring for certain constructed response items Together, these innovations account for cost savings of about $3 to $4 per student Likely to account for more as efficiencies are developed in programming and using technology for these purposes Two approaches to the use of teacher-moderated scoring. Teacher-moderated scoring can net both substantial cost reductions as well as potential professional development benefits. We used two different models for teacher-moderated scoring

Cost Reduction Strategies Two different models for teacher-moderated scoring: Professional development model - no additional teacher compensation beyond that supported by the state or district for normal professional development days (NY Regents) Stipend model - assume a $125/day stipend for teachers to score the performance items. Note: teachers were assumed to score all performance items in a distributed scoring model These strategies for using teachers as scorers reduce costs by an additional $10 to $20 per pupil (depending on whether teachers are engaged as part of professional development or are paid) Adopting all cost reduction strategies while paying teachers a $125/day stipend to score all performance tasks results in an assessment cost of $21 June 23, 2010 Assessment Solutions Group www.assessmentgroup.org

Consortia Size How big do you have to be? Stanford/Nellie Mae study found that 80% of the cost benefits of joining a consortium are realized at the 10 state size. Rough estimate is that a 5 state consortium could achieve 75%+ of the cost savings of a 10 state consortia Perhaps $3 - $6 per student $2.7 M/year for the average sized state (600K students) June 23, 2010

Where are the Cost Savings? Big cost savings opportunity in development Largely a fixed cost function Increase in forms cost partially offsets the savings Other fixed cost functions such as IT, Quality Assurance and Psychometrics provide savings Even functions that are largely variable in nature also have a fixed cost component Some functions like program management allow for economies of scale June 23, 2010 www.assessmentgroup.org

Assessment Costs by Consortium Size June 23, 2010

Where are the Cost Savings? Consortia size can make assessment technology more affordable Online test delivery (CBT and CAT) Artificial intelligence scoring of CRs More states/students more bargaining power A common assessment with common standards and operational methods s/b more efficient Need to weigh this against potential additional collaboration costs and risks June 23, 2010

PARCC & SBAC Support We recently assisted both consortia in preparing their cost estimates for the NIA responses Both consortia had innovative ideas for new assessments and a wide variety of design and operational decisions to make Each idea/design choice came with unique cost implications June 23, 2010 www.assessmentgroup.org

PARCC & SBAC Support Initially, each consortia’s design was deemed too expensive in both the operational and ongoing periods. Each needed adjustments: The number of choices and variables can be daunting as there are many variables and moving parts Ultimately, each consortia created innovative assessment systems with the designs they wanted June 23, 2010

Assessment Design Decision Tree Delivery Method Paper based Computer (linear or CAT) Mixed (both CBT and PPT) Assessment Types Summative, through course summative Interim/benchmark, End of Course, Formative Domains, special populations Indicates a major cost element for either PARCC or SBAC June 23, 2010

Decisions and Cost Variables (cont.) Development Types of items (SR, CR, Computer enhanced, PE, PT) Mix of item types Number of forms, CAT algorithm (750-1000 items per grade), number of attempts Release rates (by item type) Breach form (develop?, print?) Grades/domains tested Item bank development June 23, 2010

Decisions and Cost Variables Paper based testing/cutover to CBT How long to cut over (operating in both modes is very expensive)? Different production strategies Minimize print page “signatures” Use of color (B/W, grey scale, 4 color) Breach form (print?) Security measures (# of forms, labels, seals, student ID) June 23, 2010 www.assessmentgroup.org

Decisions and Cost Variables Logistics Transportation mode (ground, air) Carrier selection Ship from/to locations (consolidated shipping) Meetings and Travel (online vs. live) Scoring Computer vs. Human (incl. scanning and editing) June 23, 2010

Design Decisions & Costs Scoring (cont.) Human Method (teacher or 3rd party) Holistic vs. analytic scoring Requires a lot of work to develop innovative items that can be scored in a timely manner Alternatively, a test design where these items are scored during a classroom period may make sense (PEs) AI scoring for open ended items Math vs. ELA Items requiring inference can’t easily be scored using AI System training fees (fixed cost); per score costs June 23, 2010

Design Decisions & Costs Open-Ended Scoring (cont.) Double scoring/Read behind rates (by grade) Distributed vs. on-site Reporting Paper vs. online reporting Number and complexity of reports June 23, 2010 www.assessmentgroup.org

Conclusion Even a small consortium of states can achieve significant reductions in assessment cost Such a strategy can be useful in developing a new, high quality assessment or maintaining a current one during times of budgetary stress Participating in a consortium also allows for the implementation of innovative technologies that can improve assessment quality and reduce costs Teacher scoring of open-ended items is critical for implementing a high quality assessment There are a myriad of design and operational decisions that have significant cost impacts

Conclusion “You can’t always get what you want; but if you try sometime you just might find you get what you need.” - Mick Jagger www.assessmentgroup.org

Questions? Barry Topol btopol@assessmentgroup.org John Olson jolson@assessmentgroup.org Ed Roeber eroeber@assessmentgroup.org June 23, 2010 www.assessmentgroup.org