What does exam validity really mean? Andrew Martin Purdue Pesticide Programs.

Slides:



Advertisements
Similar presentations
Elliott / October Understanding the Construct to be Assessed Stephen N. Elliott, PhD Learning Science Institute & Dept. of Special Education Vanderbilt.
Advertisements

Conceptualization and Measurement
Cal State Northridge Psy 427 Andrew Ainsworth PhD
The Research Consumer Evaluates Measurement Reliability and Validity
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
1 Content-based Interpretations of Test Scores Michael Kane National Conference of Bar Examiners Maryland Assessment Research Center for Education Success.
CHAPTER FOURTEEN Testing McGraw-Hill/Irwin © 2004 The McGraw-Hill Companies, Inc., All Rights Reserved The Constitution will always protect an.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Reliability and Validity of Research Instruments
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
RESEARCH METHODS Lecture 18
USING AND PROMOTING REFLECTIVE JUDGMENT AS STUDENT LEADERS ON CAMPUS Patricia M. King, Professor Higher Education, University of Michigan.
1 An Introduction to Validity Arguments for Alternate Assessments Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland.
Chapter 4 Validity.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
Lecture 7 Psyc 300A. Measurement Operational definitions should accurately reflect underlying variables and constructs When scores are influenced by other.
1 Writing the Research Proposal Researchers communicate: Plans, Methods, Thoughts, and Objectives for others to read discuss, and act upon.
Education 795 Class Notes Factor Analysis II Note set 7.
Validity of Selection. Objectives Define Validity Relation between Reliability and Validity Types of Validity Strategies.
Classroom Assessment A Practical Guide for Educators by Craig A
Understanding Validity for Teachers
Chapter 4. Validity: Does the test cover what we are told (or believe)
VALIDITY. Validity is an important characteristic of a scientific instrument. The term validity denotes the scientific utility of a measuring instrument,
Technical Issues Two concerns Validity Reliability
Validity and Reliability
Reliability, Validity, & Scaling
Measurement in Exercise and Sport Psychology Research EPHE 348.
Ch 6 Validity of Instrument
Reliability and Validity what is measured and how well.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Technical Adequacy Session One Part Three.
Test item analysis: When are statistics a good thing? Andrew Martin Purdue Pesticide Programs.
Reliability & Validity
Introduction to Validity
Assessing Learning for Students with Disabilities Tom Haladyna Arizona State University.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Chapter 8 Validity and Reliability. Validity How well can you defend the measure? –Face V –Content V –Criterion-related V –Construct V.
Thomson South-Western Wagner & Hollenbeck 5e 1 Chapter Sixteen Critical Thinking And Continuous Learning.
Session 4 Reliability and Validity. Validity What does the instrument measure and How well does it measure what it is supposed to measure? Is there enough.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MEASUREMENT. MeasurementThe assignment of numbers to observed phenomena according to certain rules. Rules of CorrespondenceDefines measurement in a given.
MOI UNIVERSITY SCHOOL OF BUSINESS AND ECONOMICS CONCEPT MEASUREMENT, SCALING, VALIDITY AND RELIABILITY BY MUGAMBI G.K. M’NCHEBERE EMBA NAIROBI RESEARCH.
Validity: Introduction. Reliability and Validity Reliability Low High Validity Low High.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
The Development and Validation of the Evaluation Involvement Scale for Use in Multi-site Evaluations Stacie A. ToalUniversity of Minnesota Why Validate.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
Validity Evaluation NCSA Presentation NAAC/NCIEA GSEG CONSTORIA.
Validity & Reliability. OBJECTIVES Define validity and reliability Understand the purpose for needing valid and reliable measures Know the most utilized.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Language Assessment Lecture 7 Validity & Reliability Instructor: Dr. Tung-hsien He
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Consistency and Meaningfulness Ensuring all efforts have been made to establish the internal validity of an experiment is an important task, but it is.
Lecture 5 Validity and Reliability
Introduction to the Validation Phase
Reliability and Validity in Research
Tests and Measurements: Reliability
Validity.
Week 3 Class Discussion.
Command terms with definitions.
IB Vocabulary: Command Terms
Glossary Of Command Terms
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Presentation transcript:

What does exam validity really mean? Andrew Martin Purdue Pesticide Programs

What we’re going to do  Take a brief look at the evolution of validity theory from 1950 to the present.  Review modern validity theory.  Develop an analogy between test validation and pesticide enforcement.  Examine modern validity theory and its practical implications for pesticide applicator certification tests.

That’s your opinion. Everybody knows that. Our expert says…!  Sources 1. :  Cureton, E. (1951)  Cronbach, L. (1971)  Messick, S. (1989)  Shepard, L. (1993)  Kane, M. (2004)  AERA, APA, & NCME (1966, 1985, & 1999) 1. Full citations available upon request

Validity theory ca (does this sound familiar?)  It [validity] indicates how well the test serves the purpose for which it is used. Cureton (1951)

Validity defined ca  The validity of a test is an estimate of the correlation between raw test scores and the “true” (that is, perfectly reliable) criterion scores. Cureton (1951)

Problems with the criterion model  The criterion model raises two questions:  What counts as a good criterion measure?  How can the criterion measure be validated?

Validity theory ca  Validity information indicates the degree to which the test is capable of achieving certain aims…we may distinguish three of the numerous aims of testing:  Content validity  Criterion-related validity  Construct validity. APA, AERA, & NCME (1966)

How was the trinitarian model supposed to work?  Validation of an instrument calls for an integration of many types of evidence. The varieties of investigation are not alternatives any one of which would be adequate.The investigations supplement one another. Cronbach, 1971

Problems with the trinitarian model  The trinitarian model encouraged test developers to cherry-pick validity evidence.

Validity theory today  Validity is a unified concept… Messick, 1989 It is a judgment of the extent to which theory and evidence support the way test scores are interpreted and used. See also APA, AERA, & NCME, 1985 & 1999

Are there problems with a unified approach to validity?  [Current validity standards] do not help answer the question “How much evidence is enough?” Shepard, 1993

Validity and validation  Theory is supposed to guide practice.  Modern validity theory assumes a guiding role when it is coupled with a conception of validation as a practical argument.

Steps in an argument-based approach to validation  1. Specify the the proposed interpretation and use of the test scores.  2. Assemble available evidence to support the plausibility of test score interpretation and use.  3. Evaluate doubtful assumptions in detail.  4. Repeat steps 2 and 3 until the argument is plausible (or rejected) Kane, 2004

How is a validation argument like an enforcement case?  The test developer is supposed to:  gather appropriate evidence,  organize it in a compelling fashion,  and make a sound (plausible) case for how test scores will get interpreted and used.

What does this mean for me, the test developer (or how much evidence is enough)?  An argument-based approach to validation offers answers to two vexing questions:  1. How do I get started?  2. Where do I focus my resources?