Creating Assessments The three properties of good assessments.

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

Issues of Reliability, Validity and Item Analysis in Classroom Assessment by Professor Stafford A. Griffith Jamaica Teachers Association Education Conference.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Chapter 5 Measurement, Reliability and Validity.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
ASSESSMENT LITERACY PROJECT4 Student Growth Measures - SLOs.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
5.00 Understand Promotion Research  Distinguish between basic and applied research (i.e., generation of knowledge vs. solving a specific.
Copyright © Allyn & Bacon (2007) Data and the Nature of Measurement Graziano and Raulin Research Methods: Chapter 4 This multimedia product and its contents.
Assessment: Reliability, Validity, and Absence of bias
VALIDITY.
Item Response Theory. Shortcomings of Classical True Score Model Sample dependence Limitation to the specific test situation. Dependence on the parallel.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Classroom Assessment A Practical Guide for Educators by Craig A
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Understanding Validity for Teachers
Classroom Assessment A Practical Guide for Educators by Craig A. Mertler Chapter 9 Subjective Test Items.
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Technical Issues Two concerns Validity Reliability
Please check, just in case…. Announcements 1.Independent work session on the standardized test critiques next week. Meet Brigid at the Educational Diagnostician.
Measurement and Data Quality
GUIDELINES FOR SETTING A GOOD QUESTION PAPER
CHAPTER 4 Research in Psychology: Methods & Design
1 Development of Valid and Reliable Case Studies for Teaching, Diagnostic Reasoning, and Other Purposes Margaret Lunney, RN, PhD Professor College of.
Measurement in Exercise and Sport Psychology Research EPHE 348.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Technical Adequacy Session One Part Three.
Induction to assessing student learning Mr. Howard Sou Session 2 August 2014 Federation for Self-financing Tertiary Education 1.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Thoughts on Model Validation for Engineering Design George A. Hazelrigg.
Student assessment AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
Chap. 2 Principles of Language Assessment
Validity Is the Test Appropriate, Useful, and Meaningful?
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
Selecting a Sample. Sampling Select participants for study Select participants for study Must represent a larger group Must represent a larger group Picked.
The Teaching Process. Problem/condition Analyze Design Develop Implement Evaluate.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Assessment and Testing
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Validity Validity is an overall evaluation that supports the intended interpretations, use, in consequences of the obtained scores. (McMillan 17)
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Chapter 6 - Standardized Measurement and Assessment
Language Assessment. Evaluation: The broadest term; looking at all factors that influence the learning process (syllabus, materials, learner achievements,
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Survey Design Class 02.  It is a true measure  Measurement Validity is the degree of fit between a construct and indicators of it.  It refers to how.
Lesson 3 Measurement and Scaling. Case: “What is performance?” brandesign.co.za.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
© 2009 Pearson Prentice Hall, Salkind. Chapter 5 Measurement, Reliability and Validity.
Assessment in Education ~ What teachers need to know.
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
The pre- post assessment 3 points1 point Introduction Purpose Learning targets Instruments Eliminate bias & distortion Writing conventions.
Consistency and Meaningfulness Ensuring all efforts have been made to establish the internal validity of an experiment is an important task, but it is.
Language Assessment.
Data and the Nature of Measurement
Ch. 15 S. 1 What Are Psychological Tests?
Reliability and Validity in Research
Validity and Reliability
Reliability & Validity
Classroom Assessment Validity And Bias in Assessment.
Validity.
Week 3 Class Discussion.
Chapter 4 Characteristics of a Good Test
Validating Growth Models
Presentation transcript:

Creating Assessments The three properties of good assessments

Validity: A case study using The Shakespeare Test “To ensure that my lessons were effective, I chose to implement a pre- and post-test on the three-class periods that would be participating in my Shakespeare unit... After administering the pre-test, I found that approximately 90% of the students scored an ‘F’ and the other 10% landed just above the “F” grade level.”

Validity: A case study using The Shakespeare Test “At the end of the 20-day unit, I administered the post-test. My assumption was that the majority of them would score high on the test. Fortunately, my assumption paid off and I found that there were zero “D” level students and nearly 80% of the students were in the “B” to “A” range. Much like the pre-test, I found this extremely satisfying since I felt I had a few complications here and there throughout the unit.”

Creating Assessments All good assessments have three key features: All good assessments have three key features: –Validity –Reliability –Usability

Validity val·id adj. 1. Well grounded; just. 2. Producing the desired results; efficacious. 3. Having legal force; effective or binding. 4. Logic. a. Containing premises from which the conclusion may logically be derived. b. Correctly inferred or deduced from a premise. (American Heritage Dictionary)

Validity

Validity Validity: an evaluation of the adequacy and appropriateness of the interpretations and uses of assessment results. (Linn & Gronlund, 2000). Validity: an evaluation of the adequacy and appropriateness of the interpretations and uses of assessment results. (Linn & Gronlund, 2000).

Validity Let us emphasize some important points: Let us emphasize some important points: Validity refers to the appropriateness of the interpretation of the results of an assessment procedure, not to the procedure itself. Consider the following example: Consider the following example:

Dilbert’s Medical Exam One day Dilbert was not feeling well. He went to the Doctor’s office for a check-up. While there he and the doctor had a discussion about who was smarter, doctors or engineers. Dilbert lost the argument when the doctor pointed out that doctors make more money; therefore they must be smarter. After this, the doctor informed Dilbert that his test results indicated he had a terminal condition and would die within a few weeks. This is what happened next.

Validity Four Primary Categories of Validity: Four Primary Categories of Validity: –Content –Construct –Criterion –Consequential

Validity Content Validity Content Validity –How well does the sample of assessment tasks represent the domain of tasks to be measured? “Math Content Universe” CountingGeometryAlgebra Trigonometry Calculus + - x ÷ MultivariateCalculus TopologyNumericalAnalysis

Validity Construct Validity Construct Validity –How well can performance on the assessment be explained in terms of psychological characteristics? –How does one measure effort? Love? Hate? Indifference? Intelligence? Situational Awareness?

Validity Criterion Validity Criterion Validity –How accurately does performance on the assessment predict future performance or estimate the present performance on some other valued measure called a criterion?

Validity Consequential Validity Consequential Validity –How well did use of the assessment serve the intended purpose and avoid adverse effects?

Validity Cautions to keep in mind: Cautions to keep in mind: –Validity is a unitary concept. In the gestalt tradition, the “whole” of validity is greater than the sum of its various “parts.” –Validity is always specific to some use or interpretation, and no assessment is valid for all purposes. An assessment may have high validity in some aspects, but low validity for others.

Validity Cautions to keep in mind: Cautions to keep in mind: –Though we often say “validity of the test”, it is important to recognize that what is really being validated is the interpretation and use of the test results. –Validity is a matter of degree. It exists along a continuum, the same way gray exists in the continuum between black and white. At best we can classify it as “high”, “moderate”, or “low”

Validity Cautions to keep in mind: Cautions to keep in mind: –Validity involves an overall evaluative judgment. This judgment must be based on sound assessment results that include accurate measurement. Errors in the assessment process can radically affect the interpretation of the results and the resulting evaluation.

Validity To help insure the validity of your assessments, items should be: To help insure the validity of your assessments, items should be: –Written based on the content defined in the objectives. –Assessed at the specified learning level. –Reviewed by competent colleagues.

Validity The following is a list of common mistakes that will LOWER validity: The following is a list of common mistakes that will LOWER validity: –Use too few items to assess the objective. –Provide tasks or write items that do not function as intended (e.g. ambiguous, clues that lead to answers, bias, too easy, too hard, etc.). –Arrange items from hardest to easiest.

Validity Mistakes that LOWER validity (cont): Mistakes that LOWER validity (cont): –Write items that do not match the learning level specified in the objectives. –Impose unrealistic time constraints –Provide an atmosphere that is poorly suited to the purpose (e.g. noisy, messy, malfunctioning equipment). –Provide unclear or confusing directions. –Unfair or unsystematic scoring.

Validity –It is an evil for a well-taught and well- trained student to fail in an examination. –It is an evil for an unqualified student, through some inefficiency of the test, to obtain credit in an examination. New York State Department of Education, 1906