Validating Interim Assessments

Slides:



Advertisements
Similar presentations
KATIE BUCKLEY, HARVARD UNIVERSITY SCOTT MARION, CENTER FOR ASSESSMENT NATIONAL CONFERENCE ON STUDENT ASSESSMENT (NCSA) NATIONAL HARBOR, MD JUNE 22, 2013.
Advertisements

Performance Assessment
Introduction to Creating a Balanced Assessment System Presented by: Illinois State Board of Education.
The Design and Implementation of Educator Evaluation Systems, Variability of Systems and the Role of a Theory of Action Rhode Island Lisa Foehr Rhode Island.
Student Assessment Standards
Beyond Peer Review: Developing and Validating 21st-Century Assessment Systems Is it time for an audit? Thanos Patelis Center for Assessment Presentation.
How well did the assessment task do what we wanted it to do? Janina Drazek Manager — Assessment & Comparability, QCAR Queensland Studies Authority.
1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
Classroom Assessment A Practical Guide for Educators by Craig A
Understanding Validity for Teachers
Assessment Literacy Series
Experiences and requirements in teacher professional development: Understanding teacher change Sylvia Linan-Thompson, Ph.D. The University of Texas at.
Overall Teacher Judgements
PDHPE K-6 Using the syllabus for consistency of assessment © 2006 Curriculum K-12 Directorate, NSW Department of Education and Training.
NEXT GENERATION BALANCED ASSESSMENT SYSTEMS ALIGNED TO THE CCSS Stanley Rabinowitz, Ph.D. WestEd CORE Summer Design Institute June 19,
ASSESSMENT IN EDUCATION ASSESSMENT IN EDUCATION. Copyright Keith Morrison, 2004 PERFORMANCE ASSESSMENT... Concerns direct reality rather than disconnected.
Looking At Your Assessment System: A Graphical Perspective Michigan Educational Research Association Fall Conference 2012 Monday, November 19, 2012.
ASSESSMENT OF STUDENT LEARNING Manal bait Gharim.
Interim Assessments: Do You Know What You Are Buying and Why? Scott Marion, Center for Assessment Imagine: Mathematics Assessment for Learning A Convening.
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
Module 3: Unit 1, Session 3 MODULE 3: ASSESSMENT Adolescent Literacy – Professional Development Unit 1, Session 3.
CCSSO Criteria for High-Quality Assessments Technical Issues and Practical Application of Assessment Quality Criteria.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
Designing Local Curriculum Module 5. Objective To assist district leadership facilitate the development of local curricula.
OSEP Project Directors’ Conference Washington, DC July 21, 2008 Tools for Bridging the Research to Practice Gap Mary Wagner, Ph.D. SRI International.
Understanding the Common Core State Standards and Literacy Standards.
SCOTT MARION, CENTER FOR ASSESSMENT PRESENTATION AT CCSSO NCSA AS PART OF THE SYMPOSIUM ON: STUDENT GROWTH IN THE NON-TESTED SUBJECTS AND GRADES: OPTIONS.
Building an Interim Assessment System: A Workbook for School Districts CCSSO National Conference on Student Assessment Detroit, MI June 22, 2010.
Formative Interim Summative Assessments Differences, Definitions, & Applications.
Curriculum Forum Secondary Tuesday 6 June 2017
EVALUATING EPP-CREATED ASSESSMENTS
Benchmark Lesson #2: Types and Purposes of Assessments
Chapter 14 Evaluation in Healthcare Education
M-LANG project  Ref. n NO01-KA Interactive Exchange Workshop on how to use response systems and ICT tools for creating interactive learning.
American Institutes for Research
Consider Your Audience
ECML Colloquium2016 The experience of the ECML RELANG team
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
Assessment and Evaluation
Elayne Colón and Tom Dana
Classroom Assessment A Practical Guide for Educators by Craig A
Improving the Accessibility of Locally Developed Assessments CCSSO National Conference on Student Assessment 2016 Phyllis Lynch, PhD Director, Instruction,
ASSESSMENT OF STUDENT LEARNING
Consistency of Teacher Judgement
Teacher Evaluation “SLO 101”
Washback and Alternative Assessment
Illinois Performance Evaluation Advisory Council Update
Evaluating the Quality of Student Achievement Objectives
Study Questions To what extent do English language learners have opportunity to learn the subject content specified in state academic standards and.
Start with the Science & Technology Standards (2002, 2008?)
Topic Principles and Theories in Curriculum Development
Illinois Performance Evaluation Advisory Council Update
Critically Evaluating an Assessment Task
Assessments: Beyond the Claims
Unit 7: Instructional Communication and Technology
Quality in formative assessment
Assessing Academic Programs at IPFW
CLOSING As we have conducted focus groups, webinars, road shows, and even at our launch event in Austin we have heard from many educators across the state.
Assessment Literacy: Test Purpose and Use
IDEA Student Ratings of Instruction
Assessment Practices in a Balanced Assessment System
AACC Mini Conference June 8-9, 2011
Designing Your Performance Task Assessment
Assessing Students With Disabilities: IDEA and NCLB Working Together
Providing feedback to learners
Why do we assess?.
Elementary Mathematics
Aligning curriculum,instruction, and assessemnt
Presentation transcript:

Validating Interim Assessments Some Comments on Validating Interim Assessments The quality of the information and the perspectives offered by the presenters is such that I cannot offer any critiques as a discussion. Rather, what I will try to do is synthesize the points and information made. It won’t be perfect and I encourage the presenters to add/correct me and what I say… Presentation at CCSSO’s National Conference on Student Assessment Austin, TX June 30, 2017 Presenter : Thanos Patelis, HumRRO

Overview Assertions Remind us of the definition of interim assessments & types of uses Frameworks for validation and quality Recap and reinforce the presenters’ information/comments So, more concretely, I will

Assertions (Declarations, contentions, claims) Not that validating summative assessments is easy, but validating interim assessments is hard work and requires various types of evidence. But, it’s fundamental to assessment. I, consistent with frameworks about the quality of assessments, had considered validity as a component of what represents quality. However, if you consider the various types of evidence needed (as indicated by the presenters), these types of validity evidence represent all the components associated with frameworks for evaluating the quality of assessments. Fundamentally, when we are gathering evidence to support the claims of using interim assessments for instructional purposes, we are evaluating whether they are assessing rather than evaluating the assessments. Validation starts with the conceptualization of the construct and the specification of the learning targets. Not only important to do it, but also how explicit your are. I want to start with some declarations, contentions, claims that can to a large extent be found in what the presenters offered. Assessing (the process) versus the assessment (instrument)

Defining Interim Assessment Evaluate students’ knowledge and skills relative to a specific set of academic goals, typically within a limited time frame. Designed to inform decisions at both the classroom and beyond the classroom level, such as the school or district level. Administered more frequently than summative assessments. The scope and duration are more than summative and less than formative classroom assessments. Synonyms: Interim; benchmark By definition interim assessments are intended to measure a set of academic goal within a time frame (typically). Pragmatic definition to situate interim between summative and formative classroom assessments in terms of the cycle/frequency/duration. Perie, Marion, Gong, & Wurtzel, 2007

Uses of Interim Assessments Instructional – inform learning and teaching to understand and act Program Evaluation – by teacher (for example) to evaluate curriculum over repeated instructional cycles. Predictive – inform future performance Summative for Grading Professional Development - by teacher (for example) to improve own teaching and curriculum over repeated instructional cycles (sections, years) These are articulated purposes of interim assessments. Perie, Marion, & Gong, 2009

Validation “Validity refers to the degree to which evidence and theory support the interpretations of test scores for proposed uses of tests. Validity is, therefore, the most fundamental consideration in developing tests and evaluating tests… It is the interpretations of test scores for proposed uses that are evaluated, not the test itself.” (American Educational Research Association, American Psychological Association, & National Council on Measurement in Education, 2014, p.11). Kane (2013) indicated that validation consists of constructing an interpretive argument addressing four aspects of scoring, generalization, extrapolation, and implication/decision. Using this interpretive argument approach and incorporating the context permits us not only to have validation evidence, but also for demonstrating a quality interim assessment. As suggested by the presenters, this is done by the collection of corroborating evidence across multiple areas. Validation is evidence Compilation of evidence to substantiate the meaning of scores The evidence represents various components in multiple areas (as suggested by the presenters) In this you cannot ignore the context and doing this cannot be done without knowing what it is that you are evaluating (i.e., the content/construct/standards/learning outcomes/learning progressions)

Criteria for Evaluating Quality: Alignment - Standards and assessments are aligned Diagnostic Value - Multiple item types are used to increase diagnostic value for instructional planning Fairness - Assessments are fair for all students including English language learners and students with disabilities Technical - Assessments show quality of test reliability and validity Utility - User-friendly results and guidance on interpreting and using results to improve instruction are provided Feasibility - Assessments are feasible and worth of time and money investment by schools and districts Components for evaluating the quality (as you can see) include validity as a component. But these components are things that the presenters (to varying degrees) indicated are the evidence gathering for validity! Herman & Baker, 2005

Criteria for Evaluating Depend on Purpose: Instructional Purpose: Fit with instruction and represent opportunity to learn Assessment system has improved student learning based on rigorous research Evidence that score reports facilitate meaningful and useful instructional interpretations Guidelines provided on how results should inform instructional decisions Each part must link closely to curricular goals Scope should be such that instruction can occur Type of question should provide useful information about students’ understanding and cognition and include open-ended, as well as multiple-choice questions Assessment should measure instructional and curricular goals and provide information showing students’ in-depth understanding Others have suggested that the information gathered to evaluate quality is contingent on the purpose. So, here is a list of aspects that are evaluating the quality of interim assessments associated with instructional purpose. Perie, Marion, & Gong, 2009

In validating interim assessments… Specify purpose, use, and the construct (learning objectives, progressions, etc.) Specificity is a must! Know context Instructional Teacher Students Gather evidence and information. Taxonomy of effort: Match Corroborate Compare Replicate Assess usefulness So, here are the results of my distillation of the evidence. Three components of the content, the context, and the activities. Content – learning progressions Context – reference to Paul Nichols’ framework in defining assessments by representing the domain, but also the teaching model.

Components of Validation: Alignment – Are standards, learning objectives (progressions) and assessments aligned? (Match) Diagnostic Value – Corroborate the information gathered with other sources Fairness – Compare performance for all students and by group including English language learners and students with disabilities Technical – Replicate results over time, across groups (classes) Utility – Are results used to improve instruction? Feasibility – Can the assessments be used by teachers including results? Here is the framework for evaluating the quality of interim assessments, tweaked to represent the approach/components of the validation process.

Synthesis The conceptual basis of the assessments must be developed explicitly. There are multiple components of evidence needed. The validation (if utilize notion of a validity argument) represents the components involved in ensuring and evaluating the quality of interim assessment. Validation involves alignment, corroborating evidence, comparisons, replications, and ascertaining usefulness. Don’t forget the interim assessment users both in terms of the utility and feasibility of the assessments for them, but also in their assessment knowledge and skills. When packaging the information representing all the various components, it must be clearly articulated by organizing the information in a logical way and clearly communicating the process and results. Guidelines and standards should incorporate some of the specificity in the validation process and evidence. In summary, I offer some statements that the presenters indicated. I apologize for the perceive simplicity of these statements or if they are just obvious, but they are fundamental and sometimes (maybe often) neglected. Learning progressions is an important fundamental component. You cannot assess or even align the assessment and the results to something without a map.

References American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (2014). Standards for Educational and Psychological Testing. Washington, DC: American Educational Research Association. Herman, J. L. & Baker, E. L. (2005). Making benchmark testing work. Educational Leadership, 63(3), 48-54. Kane, M. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50(1), 1-73. Perie, M., Marion, S., & Gong, B. (2009). Moving toward a comprehensive assessment system: A framework for considering interim assessment. Educational Measurement: Issues and Practice, 28(3), 5-13. Perie, M., Marion, S., Gong, B., & Wurtzel, J. (2007). The role of interim assessments in a comprehensive assessment system. Washington, DC: The Aspen Institute.

Questions? Contact: Thanos Patelis tpatelis@humrro.org