Questions What are the sources of error in measurement?

Slides:



Advertisements
Similar presentations
Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Advertisements

Chapter 8 Flashcards.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Topics: Quality of Measurements
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
Reliability Definition: The stability or consistency of a test. Assumption: True score = obtained score +/- error Domain Sampling Model Item Domain Test.
VALIDITY AND RELIABILITY
Reliability - The extent to which a test or instrument gives consistent measurement - The strength of the relation between observed scores and true scores.
Part II Sigma Freud & Descriptive Statistics
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Measurement Reliability and Validity
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Reliability and Validity of Research Instruments
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
Test Validity: What it is, and why we care.
RELIABILITY & VALIDITY What is Reliability? What is Reliability?What is Reliability?What is Reliability? How Can We Measure Reliability? How Can We Measure.
RELIABILITY & VALIDITY
Concept of Measurement
Item PersonI1I2I3 A441 B 323 C 232 D 112 Item I1I2I3 A(h)110 B(h)110 C(l)011 D(l)000 Item Variance: Rank ordering of individuals. P*Q for dichotomous items.
Measurement: Reliability and Validity For a measure to be useful, it must be both reliable and valid Reliable = consistent in producing the same results.
Validity, Reliability, & Sampling
Measurement Joseph Stevens, Ph.D. ©  Measurement Process of assigning quantitative or qualitative descriptions to some attribute Operational Definitions.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Psychometrics Timothy A. Steenbergh and Christopher J. Devers Indiana Wesleyan University.
Measurement and Data Quality
Reliability, Validity, & Scaling
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
Reliability and Validity what is measured and how well.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Reliability & Validity
Validity Is the Test Appropriate, Useful, and Meaningful?
Tests and Measurements Intersession 2006.
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Measurement Validity.
EPSY 546: LECTURE 3 GENERALIZABILITY THEORY AND VALIDITY
Chapter 8 Validity and Reliability. Validity How well can you defend the measure? –Face V –Content V –Criterion-related V –Construct V.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
CHAPTER OVERVIEW The Measurement Process Levels of Measurement Reliability and Validity: Why They Are Very, Very Important A Conceptual Definition of Reliability.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MEASUREMENT. MeasurementThe assignment of numbers to observed phenomena according to certain rules. Rules of CorrespondenceDefines measurement in a given.
MOI UNIVERSITY SCHOOL OF BUSINESS AND ECONOMICS CONCEPT MEASUREMENT, SCALING, VALIDITY AND RELIABILITY BY MUGAMBI G.K. M’NCHEBERE EMBA NAIROBI RESEARCH.
Reliability & Validity  Reliability  “dependability”  is the indicator consistent?  same result every time?  Does not necessary measure what you think.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Chapter 9 Correlation, Validity and Reliability. Nature of Correlation Association – an attempt to describe or understand Not causal –However, many people.
Spring 2015 Kyle Stephenson
Measurement MANA 4328 Dr. Jeanne Michalski
2. Main Test Theories: The Classical Test Theory (CTT) Psychometrics. 2011/12. Group A (English)
Measuring Research Variables
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
1 Measurement Error All systematic effects acting to bias recorded results: -- Unclear Questions -- Ambiguous Questions -- Unclear Instructions -- Socially-acceptable.
MGMT 588 Research Methods for Business Studies
Reliability and Validity
Concept of Test Validity
Test Validity.
CHAPTER 5 MEASUREMENT CONCEPTS © 2007 The McGraw-Hill Companies, Inc.
Tests and Measurements: Reliability
Reliability & Validity
Introduction to Measurement
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
پرسشنامه کارگاه.
5. Reliability and Validity
Reliability and Validity of Measurement
Reliability & Validity
Reliability & Validity
Measurement Concepts and scale evaluation
Presentation transcript:

Questions What are the sources of error in measurement? How are these sources related to one another? How are these sources related to the ideas of reliability and validity? What are easier to deal with: problems in reliability or problems in validity? How is Cronbach’s alpha related to other measures of reliability? Why is it the disciplinary standard? What are the components of Cronbach’s reliability? From the point of view of generalizability theory

Generalizability Theory Matt, 2006

Reliability – formal definition Total variance in scores Systematic differences between subjects Unsystematic differences between subjects due to internal (to subjects) or external measurement errors

Example You measured the mathematics aptitude of a child who is suspected to be dysnumeric. The Wechsler standard aptitude score was 68. The reliability of the aptitude test was 0.88. Could you report to the parents that the child appears to be dysnumeric.

How do we increase reliability Reliability is proportional to the true score variance Larger the number of items, larger the variance, higher the reliability (under certain assumptions) # items new / # items old

“Generalizability” Mean sq person = estimate of the variance of all possible persons Mean sq item = estimate of the variance of all possible items Mean sq error = estimate of the unexplained variance

“Generalizability” Mean sq person = estimate of the variance of all possible persons Mean sq item = estimate of the variance of all possible items Mean sq error = estimate of the unexplained variance TRUE person-to-person variance RELIABILITY = TOTAL person-to-person variance MS (person) – MS (error) RELIABILITY = MS (person)

Generalizability Theory Data consist of person-item scores Variability in the data: Variability across individuals Variability across items Variability across person*item combinations

SPSS output

Generalizability with multiple “facets” Persons who are assessed Items used Informants providing data Occasions of measurement Settings in which measures were taken

Validity

Validity Reliable Valid Not Reliable Not Valid Systematic errors Unsystematic errors RELIABILITY Reliable Valid Not Reliable Not Valid

Types of Validity Content validity Criterion validity Concurrent validity Predictive validity Construct validity Convergent validity Discriminant validity

Content validity The test should elicit a range of responses that are representative of the entire domain of relevant attributes (e.g., skills, emotions, behaviors, attitudes, traits, symptoms). Achievement tests: Are all the objectives of instruction represented? Other assessments: Use guidance from theory and experts

Criterion Validity The test results should be in agreement with other accepted measures of the same attribute.

Reasons for criterion validity to be biased Group differences (Aiken unclear) Subgroups will have lower criterion validity than the aggregate due to restriction of range problem. If a test is developed and validated on a group, cross validation estimates will be lower than the original estimates of criterion validity. Length of the test Longer the test, higher the variance, higher the covariance, higher the criterion validity Criterion contamination When validation is not blind, criterion validity estimates will be upward biased

Construct validity Extent to which an instrument measures the targeted construct. Especially relevant for personality tests, attitude tests, and assessments of psychological problems/disorders. Establishing construct validity Expert judgment Reliability analysis Group differences should be in the expected direction. Convergent and discriminant validity (varying measures and methods) Talk-through administration of the assessment

Validity and social consequences of tests Social consequences of tests may be undesirable Because of the test Because of the phenomenon that the test measures