1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.

Slides:



Advertisements
Similar presentations
KATIE BUCKLEY, HARVARD UNIVERSITY SCOTT MARION, CENTER FOR ASSESSMENT NATIONAL CONFERENCE ON STUDENT ASSESSMENT (NCSA) NATIONAL HARBOR, MD JUNE 22, 2013.
Advertisements

The Network of Dynamic Learning Communities C 107 F N Increasing Rigor February 5, 2011.
Validity in Action: State Assessment Validity Evidence for Compliance with NCLB William D. Schafer, Joyce Wang, and Vivian Wang University of Maryland.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Lessons Learned from AYP Decision Appeals Prepared for the American Educational Research Association Indiana Department of Education April 15, 2004.
Alternate Assessments on Alternate Achievement Standards Student Population Jacqueline F. Kearns, Ed.D. Elizabeth Towles-Reeves, MS.
1 An Introduction to Validity Arguments for Alternate Assessments Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland.
Assessing for Deep Understanding A Guide for Learning Sciences Researchers A translation of academic writing Web-Based Learning Designers.
MCAS-Alt: Alternate Assessment in Massachusetts Technical Challenges and Approaches to Validity Daniel J. Wiener, Administrator of Inclusive Assessment.
CLOSING THOUGHTS The long and winding road of alternate assessments Where we started, where we are now, and the road ahead! Rachel F. Quenemoen, Senior.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Standard Setting Inclusive Assessment Seminar Marianne.
Setting Alternate Achievement Standards Prepared by Sue Rigney U.S. Department of Education NCEO Teleconference March 21, 2005.
Consistency/Reliability
NATIONAL CENTER ON EDUCATIONAL OUTCOMES University of Minnesota Rachel Quenemoen Cammy Lehr Martha Thurlow.
2004 CCSSO Large-scale Conference Peasley, Deeter, Quenemoen Measurement Purgatory or Best Practice? Alternate Assessment for Students with Significant.
Assessment Population and the Validity Evaluation
Principles of High Quality Assessment
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments 1 Introduction to Comparability Inclusive Assessment Seminar.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
AACC 1 Helping States and Regional Centers Meet NCLB Goals: State Accountability Systems Stanley Rabinowitz, Ph.D. WestEd CRESST Conference January 22-23,
Classroom Assessment A Practical Guide for Educators by Craig A
Understanding Validity for Teachers
Assessment of Special Education Students
Challenges in Developing a University Admissions Test & a National Assessment A Presentation at the Conference On University & Test Development in Central.
ECD in the Scenario-Based GED ® Science Test Kevin McCarthy Dennis Fulkerson Science Content Specialists CCSSO NCSA June 29, 2012 Minneapolis This material.
ERIKA HALL CENTER FOR ASSESSMENT PRESENTATION AT THE 2014 NATIONAL CONFERENCE ON STUDENT ASSESSMENT NEW ORLEANS JUNE 25, 2014 The Role of a Theory of Action.
Psychometric Issues in the Use of Testing Accommodations Chapter 4 David Goh.
Focus on Learning: Student Outcomes Assessment and the Learning College.
Criteria for Procuring and Evaluating High-Quality Assessments: A State Perspective National Conference on Student Assessment New Orleans, Louisiana June.
Monitoring through Walk-Throughs Participants are expected to purpose the book: The Three-Minute Classroom Walk-Through: Changing School Supervisory.
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
Committee on the Assessment of K-12 Science Proficiency Board on Testing and Assessment and Board on Science Education National Academy of Sciences.
Student Learning Objectives: Approval Criteria and Data Tracking September 17, 2013 This presentation contains copyrighted material used under the educational.
Rationale for the Evaluation Frameworks Project Previous Paradigm in Our State A Compliance Exercise, Often Based Solely on Positional Authority.
{ Principal Leadership Evaluation. The VAL-ED Vision… The construction of valid, reliable, unbiased, accurate, and useful reporting of results Summative.
Visions for the Future: Inclusive Assessments Jacqueline F. Kearns, Ed.D. University of Kentucky.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
1. Housekeeping Items June 8 th and 9 th put on calendar for 2 nd round of Iowa Core ***Shenandoah participants*** Module 6 training on March 24 th will.
Illustration of a Validity Argument for Two Alternate Assessment Approaches Presentation at the OSEP Project Directors’ Conference Steve Ferrara American.
Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University.
Enhancing the Technical Quality of the North Carolina Testing Program: An Overview of Current Research Studies Nadine McBride, NCDPI Melinda Taylor, NCDPI.
The Why (Waiver & Strategic Plan) Aligned to research: MET Study Components: Framework/Multiple Measures Pilot Requirements Timeline.
1 Comprehensive Accountability Systems: A Framework for Evaluation Kerry Englert, Ph.D. Paper Presented at the Canadian Evaluation Society June 2, 2003.
CAROLE GALLAGHER, PHD. CCSSO NATIONAL CONFERENCE ON STUDENT ASSESSMENT JUNE 26, 2015 Reporting Assessment Results in Times of Change:
1% + 2% = ______________: ADDING UP WHAT WE KNOW & DON’T KNOW ABOUT ALTERNATE ASSESSMENTS Stephen N. Elliott, PhD Gerald Tindal, PhD Vanderbilt UniversityUniversity.
Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment.
1 Evaluating the Validity of State Accountability Systems: Examples of Evaluation Studies Scott Marion Center for Assessment Presented at AERA, Division.
Assessment Design. Four Professional Learning Modules 1.Unpacking the AC achievement standards 2.Validity and reliability of assessments 3. Confirming.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Data Report July Collect and analyze RtI data Determine effectiveness of RtI in South Dakota in Guide.
Possible Evaluation Model: Reactions from the Field David T. Conley, Ph.D. Professor, University of Oregon.
Slide 1, Moving From Challenge To Action: Accountability Supporting Student Learning Joan L. Herman UCLA Graduate School of Education & Information.
Alternative Assessment Chapter 8 David Goh. Factors Increasing Awareness and Development of Alternative Assessment Educational reform movement Goals 2000,
A presentation of methods and selected results 1.
Validity Evaluation NCSA Presentation NAAC/NCIEA GSEG CONSTORIA.
CTB CADDS Sally Valenzuela Director, Publishing Strategic Initiatives CTB/McGraw-Hill.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
SCOTT MARION, CENTER FOR ASSESSMENT PRESENTATION AT CCSSO NCSA AS PART OF THE SYMPOSIUM ON: STUDENT GROWTH IN THE NON-TESTED SUBJECTS AND GRADES: OPTIONS.
CDIO: Overview, Standards, and Processes (Part 2) Doris R. Brodeur, November 2005.
Next Generation Iowa Assessments
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Consequential Validity
Validity and Reliability
Week 3 Class Discussion.
Federal Policy & Statewide Assessments for Students with Disabilities
Validating Interim Assessments
Assessment Population and the Validity Evaluation
William D. Schafer, Joyce Wang, and Vivian Wang University of Maryland
Assessment Literacy: Test Purpose and Use
Presentation transcript:

1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October 11-12, 2007

Marion. Center for Assessment. MARCES Key Points Evaluating the technical quality of AA- AAS must focus on the validity of the test based inferences Specifying the interpretative argument will help create the validity evaluation plan –Consequences are essential Prioritize and get started!

Marion. Center for Assessment. MARCES It’s About Validation The purpose of technical documentation is to provide data for critically evaluating the inferences from the AA-AAS scores and the logic of the interpretative argument

Marion. Center for Assessment. MARCES What about alignment, reliability, comparability, etc? Reliability and comparability might be overrated, especially in the narrow way that reliability has been applied in AA-AAS evaluations Alignment, if pursued at a deep level, can provide useful evidence about the nature of the construct

Marion. Center for Assessment. MARCES Specifying the argument Simply saying that the AA-AAS is being designed to fulfill NCLB and IDEA requirements is inadequate Like a theory of action, the interpretative argument makes the values explicit and specifies the logical connections among the various components of the system, for example: –observed scores will increase for students in response to high quality instruction –the quality of school level instruction will increase as a result of appropriate use of test scores to target professional development –as a result of decisions based on AA-AAS scores, the educational opportunities of students will improve These all lead to important validity inquires

Marion. Center for Assessment. MARCES Consequences are Central With AA-AAS we are trying to balance technical rigor with social justice There is a fundamental belief that educational opportunities for students with significant cognitive disabilities will improve when included in test based accountability systems –If this is a value of the system—and I believe it is for most state AA-AAS systems—than the evaluation of technical quality must include a critical examination of the intended positive and unintended negative effects of the uses of the test scores in this system

Marion. Center for Assessment. MARCES Prioritizing and Focusing Several authors, including Kane (2006), Haertel (1999), Lane (2003), Ryan (2002), and Shepard (1993) have offered suggestions for prioritizing and focusing validity evaluations –There is no “right” way, but many wrong ways States need to develop a validity studies plan, based upon the interpretative argument (or other legitimate organizing framework) to organize both the studies and the synthesis of evidence –Many of these studies—particularly consequential studies—need early planning and some initial data collection

OBSERVATION INTERPRETATION COGNITION  Student Population  Academic Content  Theory of Learning  Assessment System  Test Development  Administration  Scoring  Reporting  Alignment  Item Analysis/DIF/Bias  Measurement Error  Scaling and Equating  Standard Setting VALIDITY EVALUATION  Empirical Evidence  Theory and Logic (argument)  Consequential Features A heuristic to help organize and focus the validity evaluation (Marion, Quenemoen, & Kearns, 2006)