ARG symposium discussion Dylan Wiliam Annual conference of the British Educational Research Association; London, UK: 2007 www.dylanwiliam.net.

Slides:



Advertisements
Similar presentations
School Based Assessment and Reporting Unit Curriculum Directorate
Advertisements

Wynne Harlen. What do you mean by assessment? Is there assessment when: 1. A teacher asks pupils questions to find out what ideas they have about a topic.
Assessment Systems for the Future: the place of assessment by teachers A project of the Assessment Reform Group, funded by the Nuffield Foundation.
The role of teachers in the assessment of learning
Formative assessment and contingency in the regulation of learning processes Contribution to a Symposium entitled “Toward a theory of classroom assessment.
Summative Assessment Kansas State Department of Education ASSESSMENT LITERACY PROJECT1.
Utah Effective Teaching Standards-based Jordan Performance Appraisal System Orientation (UETS-based JPAS)
The Research Consumer Evaluates Measurement Reliability and Validity
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
ASSESSMENT LITERACY PROJECT4 Student Growth Measures - SLOs.
Systematic Assessment of Student Achievement February 5, 2007.
A Terse Self-Test about Testing
Grading. Why do we grade? To communicate To tell students how they are doing To tell parents how students are doing To make students uneasy To wield power.
The search for the ‘dark matter’ of teacher quality Dylan Wiliam
Using Assessment Data for Improving Teaching Practice ACER Conference August 2009 Professor Helen Timperley Faculty of Education University of.
Not invented here: the baffling insularity of assessment practices in higher education Dylan Wiliam Keynote presentation at the University.
How to Build Learning Progressions: Formative Assessment’s Basic Blueprints Presentation 3 Siobhán Leahy Dylan Wiliam.
From Evidence to Action: A Seamless Process in Formative Assessment? Margaret Heritage Jinok Kim Terry Vendlinski American Educational Research Association.
BASIC PRINCIPLES OF ASSSESSMENT RELIABILITY & VALIDITY
Assessment Primary and Secondary Probationer Teachers.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
V ALIDITY - C ONSEQUANTIALISM Assoc. Prof. Dr. Sehnaz Sahinkarakas.
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
Chapter 4. Validity: Does the test cover what we are told (or believe)
Achievement Testing Dale Pietrzak, Ed.D., LPC-MH, CCMHC University of South Dakota Counseling & Psychology in Education.
Shawna Williams BC TEAL Annual Conference May 24, 2014.
Challenges in Developing a University Admissions Test & a National Assessment A Presentation at the Conference On University & Test Development in Central.
Classroom Assessment A Practical Guide for Educators by Craig A
Six degrees of integration: an agenda for joined-up assessment Dylan Wiliam Annual Conference of the Chartered Institute of Educational.
ASSESSMENT LITERACY: A BONA FIDE “MAGIC BULLET” FOR EDUCATION W. James Popham University of California, Los Angeles California Educational Research Association.
Overall Teacher Judgements
Reflections on pedagogy Dylan Wiliam Pedagogy, Space, Place Conference November
“We will lead the nation in improving student achievement.” CLASS Keys TM Module 1: Content and Structure Spring 2010 Teacher and Leader Quality Education.
DLM Early Childhood Express Assessment in Early Childhood Dr. Rafael Lara-Alecio Dr. Beverly J. Irby
A Role for Formalized Tools in Formative Assessment Bob Dolan, Senior Research Scientist, Pearson CCSSO NCSA | National Harbor |
Stephen C. Court Educational Research and Evaluation, LLC A Presentation at the First International Conference on Instructional Sensitivity Achievement.
Teaching Today: An Introduction to Education 8th edition
1. What do you know when you know the test results? The meanings of educational assessments Annual Conference of the International Association for Educational.
Introduction to Validity
What kinds of assessment support learning of key competences? Dylan Wiliam EC seminar on the assessment of key competences Brussels, Belgium,
CT 854: Assessment and Evaluation in Science & Mathematics
Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University.
Teaching Today: An Introduction to Education 7 th edition Part 3: Teaching and Assessing Chapter 10: Assessing Learning Armstrong, Henson, & Savage Teaching.
1. Designing an assessment system Presentation to the Scottish Qualifications Authority, August 2007 Dylan Wiliam Institute of Education, University of.
Assessment Design. Four Professional Learning Modules 1.Unpacking the AC achievement standards 2.Validity and reliability of assessments 3. Confirming.
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Do Your Assessment Instruments Measure Up? Gerald Tindal UO.
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
FLAGSHIP STRATEGY 1 STUDENT LEARNING Focus on mathematics.
The Development and Validation of the Evaluation Involvement Scale for Use in Multi-site Evaluations Stacie A. ToalUniversity of Minnesota Why Validate.
1. Statistical power in educational settings Workshop at Wellcome seminar on educational research, May 2008 Dylan Wiliam Institute of Education, University.
Revisiting Target Setting: A Complex Task Judy M. Parr University of Auckland EHSAS Seminars October 2008.
ASSESSMENT: FORMATIVE & SUMMATIVE Practices for the Classroom.
Why teaching will never be a research-based profession (and why that’s a Good Thing) Dylan Wiliam 1.
Key actions:  No national regulator, but AITSL to drive improvement  Provisional and full accreditation  Clear and explicit instructions for providing.
LAPS symposium discussion Dylan Wiliam Annual conference of the British Educational Research Association; London, UK:
Assessment Small Learning Communities. The goal of all Small Learning Communities is to improve teaching, learning, and student outcomes A rigorous, coherent.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
C R E S S T / CU University of Colorado at Boulder National Center for Research on Evaluation, Standards, and Student Testing Measuring Adequate Yearly.
INTRODUCTION TO ASSESSMENT METHODS USED IN MEDICAL EDUCATION AND THEIR RATIONALE.
Implementing Formative Assessment Processes: What's Working in Schools and Why it is Working Sophie Snell & Mary Jenatscheck.
Assessment of Learning 1
MENTEP, Brussels Janet Looney
What do you know when you know the test results
CEM Primary Overview.
Designing an assessment system
Quality in formative assessment
Designing an assessment system
Week 2 Evaluation Framework
Presentation transcript:

ARG symposium discussion Dylan Wiliam Annual conference of the British Educational Research Association; London, UK:

Validity Validity is a property of inferences, not assessments No such thing as a biased assessment Validity subsumes all aspects of assessment quality Reliability Content coverage But not impact (Popham: right concern, wrong concept) “Validity is an integrative evaluative judgment of the degree to which empirical evidence and theoretical rationales support the adequacy and appropriateness of inferences and actions based on test scores or other modes of assessment” (Messick, 1989 p. 13)

Messick (1989) Result interpretation Result use Evidential basis Content validityConstruct validity/utility Consequential basis Value implicationsSocial consequences

Validity As has been stressed several times already, it is not that adverse social consequences of test use render the use invalid, but, rather, that adverse social consequences should not be attributable to any source of test invalidity such as construct-irrelevant variance. If the adverse social consequences are empirically traceable to sources of test invalidity, then the validity of the test use is jeopardized. If the social consequences cannot be so traced—or if the validation process can discount sources of test invalidity as the likely determinants, or at least render them less plausible—then the validity of the test use is not overturned. Adverse social consequences associated with valid test interpretation and use may implicate the attributes validly assessed, to be sure, as they function under the existing social conditions of the applied setting, but they are not in themselves indicative of invalidity. (Messick, 1989, p )

Koretz, Linn, Dunbar, Shepard (1991)

Sensitivity to instruction Average cohort progress: 0.3 sd per year Good teachers (+1 sd) produce 0.4 sd per year Poor teachers (-1 sd) produce 0.2 sd per year Giving all disadvantaged children above average teachers, and all advantaged children below average teachers would take 5 years to eradicate the achievement gap…

So… Although teacher quality is the single most important determinant of student progress… …the effect is small compared to the accumulated achievement over the course of a learner’s education… …inferences that school outcomes are indications of the contributions made by the school are almost certainly invalid.

Teacher involvement in high-stakes assessments In a high-stakes environment, the challenge is to produce assessments worth teaching to Assessments therefore need to be Distributed Cumulative and synoptic

Teacher professional development Effective teacher engagement in summative assessment requires new teacher knowledge Effective teacher engagement in formative assessment requires new teacher behaviours These require very different kinds of professional development, and different structures to sustain them