Characteristics of Sound Tests

Slides:



Advertisements
Similar presentations
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Advertisements

© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
Part II Sigma Freud & Descriptive Statistics
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
Part II Sigma Freud & Descriptive Statistics
MEQ Analysis. Outline Validity Validity Reliability Reliability Difficulty Index Difficulty Index Power of Discrimination Power of Discrimination.
Copyright © Allyn & Bacon (2007) Data and the Nature of Measurement Graziano and Raulin Research Methods: Chapter 4 This multimedia product and its contents.
Chapter 4 Validity.
REVIEW I Reliability Index of Reliability Theoretical correlation between observed & true scores Standard Error of Measurement Reliability measure Degree.
MEASUREMENT. Measurement “If you can’t measure it, you can’t manage it.” Bob Donath, Consultant.
Concept of Measurement
Beginning the Research Design
Introduction to Senior Fitness Test Instructor: Jessie Jones, Ph.D.
Manipulation and Measurement of Variables
Psych 231: Research Methods in Psychology
Manipulation and Measurement of Variables
EVALUATION OF DR.MOHAMMED AL NAAMI, FRCSC, FACS, M Ed. Using O bjective S tructured C linical E xamination (OSCE)
Validity, Reliability, & Sampling
Classroom Assessment A Practical Guide for Educators by Craig A
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
The Practice of Social Research
Measurement and Data Quality
Validity and Reliability
Collecting Quantitative Data
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Emily H. Wughalter, Ed.D. Measurement & Evaluation Spring 2010.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
1 SELECTION 2BC3 Week 5 ________________________ Dr. Teal McAteer DeGroote School of Business McMaster University.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Induction to assessing student learning Mr. Howard Sou Session 2 August 2014 Federation for Self-financing Tertiary Education 1.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Principles of Test Construction
Reliability REVIEW Inferential Infer sample findings to entire population Chi Square (2 nominal variables) t-test (1 nominal variable for 2 groups, 1 continuous)
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Chap. 2 Principles of Language Assessment
Chapter Five Measurement Concepts. Terms Reliability True Score Measurement Error.
Assessment What is it? Collection of relevant information for the purpose of making reliable curricular decisions and discriminations among students (Gallahue,
Measurement Validity.
Advanced Research Methods Unit 3 Reliability and Validity.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Reliability and Validity Themes in Psychology. Reliability Reliability of measurement instrument: the extent to which it gives consistent measurements.
REVIEW I Reliability scraps Index of Reliability Theoretical correlation between observed & true scores Standard Error of Measurement Reliability measure.
DENT 514: Research Methods
Chapter 6 - Standardized Measurement and Assessment
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Assessing Student Performance Characteristics of Good Assessment Instruments (c) 2007 McGraw-Hill Higher Education. All rights reserved.
Dr. Jeffrey Oescher 27 January 2014 Technical Issues  Two technical issues  Validity  Reliability.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
Data Collection Methods NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN.
Consistency and Meaningfulness Ensuring all efforts have been made to establish the internal validity of an experiment is an important task, but it is.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
by Holcomb Hathaway Publishers
Data and the Nature of Measurement
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
DUMMIES RELIABILTY AND VALIDITY FOR By: Jeremy Starkey Lijia Zhang
Associated with quantitative studies
Week 3 Class Discussion.
Measurement Concepts and scale evaluation
Presentation transcript:

Characteristics of Sound Tests Instructor: Jessie Jones, Ph.D. Co-director, Center for Successful Aging California State University, Fullerton

Criteria for Evaluating Tests Reliability Validity Discrimination Performance Standards Social Acceptability Feasibility

Test Reliability Refers to the consistency of a score from one trial to the next (especially from one day to another). test-retest reliability r = .80

Test Reliability Test objectivity- refers to the degree of accuracy in scoring a test. Also referred to as rater reliability

Rater Reliability Is especially important if measures are going to be collected on multiple occasions and/or by more than one rater. Intrarater reliability refers to the same evaluator. Interrater reliability refers to different evaluators.

Test Reliability How to increase scoring precision Practice giving the test to a sample of clients Follow the exact published protocol Provide consistent motivation Provide rest to reduce fatigue Help to reduce client fear Note any adaptations in test protocol Chair Stand

Reliability - Review Reliability Test-retest reliability Test Objectivity Intra-rater reliability Inter-rater reliability

Test Validity A valid test is one that measures what it is intended to measure. Physical fitness Functional limitations Motor and sensory impairments Fear-of-falling Tests must be validated on intended clients

Types of Validity Content Construct Criterion

Test Validity Content Validity – the degree to which a test reflects a defined “domain” of interest. Also referred to as “face” or “logical” validity. Example: Berg Balance Scale Domain of interest is balance. Participant performs a series of 14 functional tasks that require balance.

Test Validity Construct-related - the degree to which a test measures a particular construct. A construct is an attribute that exists in theory but cannot be directly observed. Example Test: 8’ Up & Go Construct measured is functional mobility

Test Validity Criterion-related – evidence demonstrates that test scores are statistically related to one or more outcome criteria. Concurrent Validity Predictive Validity

Criterion-Related Concurrent validity – the degree to which a test correlates with a criterion measure. Criterion measure is often referred to as the “gold standard” measure. > .70 Example: Chair Sit & Reach

Criterion-Related Predictive Validity evidence demonstrates the degree of accuracy with which an assessment predicts how participants will perform in a future situation.

Predictive Validity Example Test: Berg Balance Scale Older adults who score above 46/56 have a high probability of not falling when compared to older adults who score below this cutoff.

Validity - Review Content-related Construct-related Criterion-related Concurrent validity Predictive validity

Discrimination Power Important for measuring different ability levels, and measuring over time. Continuous measure tests Result in a spread of scores Avoid “ceiling effects”- test too easy Avoid “floor effects” – test too hard Responsiveness

Discrimination Power Examples: Senior Fitness Test (ratio scale) Uses time and distance measures FAB and BBS (5 pt. ordinal scale) Allows for “more change in scores” than Tinetti’s POMA; FEMBAF which only have 2-3 point scales).

Characteristics of Sound Tests Performance Standards Evaluated relative to a peer group (norm-referenced standards). Example: Senior Fitness Test In relation to predetermined, desired outcomes (criterion-referenced standards) Example: 8 ft Up & Go

Other Characteristics of Sound Tests Social acceptability – meaningful Feasibility – suitable for use in a particular setting

Review! Reliability Validity Discrimination Performance Standards Social Acceptability Feasibility