Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability.

Slides:



Advertisements
Similar presentations
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Advertisements

How good are our measurements? The last three lectures were concerned with some basics of psychological measurement: What does it mean to quantify a psychological.
Conceptualization and Measurement
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
Some (Simplified) Steps for Creating a Personality Questionnaire Generate an item pool Administer the items to a sample of people Assess the uni-dimensionality.
Reliability and Validity
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
Reliability & Validity.  Limits all inferences that can be drawn from later tests  If reliable and valid scale, can have confidence in findings  If.
Ellen Berscheid, Mark Snyder & Allen M. Omoto
Quiz Do random errors accumulate? Name 2 ways to minimize the effect of random error in your data set.
Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability.
Reliability, the Properties of Random Errors, and Composite Scores.
CH. 9 MEASUREMENT: SCALING, RELIABILITY, VALIDITY
Measurement  The process whereby individual instances within a defined population are scored on an attribute according to rules Usually given a numeric.
RESEARCH METHODS Lecture 18
Item PersonI1I2I3 A441 B 323 C 232 D 112 Item I1I2I3 A(h)110 B(h)110 C(l)011 D(l)000 Item Variance: Rank ordering of individuals. P*Q for dichotomous items.
MGTO 231 Human Resources Management Personnel selection I Dr. Kin Fai Ellick WONG.
Sampling Basics 2011, 9, 13. Last Class: Measurement  Scale of measurement –Nominal scale –Ordinal scale –Interval-ratio scale  Reliability: Free of.
Psych 231: Research Methods in Psychology
Variables cont. Psych 231: Research Methods in Psychology.
Validity, Reliability, & Sampling
Reliability & Validity the Bada & Bing of YOUR tailored survey design.
Quiz Name one latent variable Name 2 manifest variables that are indicators for the latent variable.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Internal Consistency Reliability Analysis PowerPoint.
Reliability, Validity, & Scaling
Experimental Research
Measurement in Exercise and Sport Psychology Research EPHE 348.
Student Engagement Survey Results and Analysis June 2011.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Goals for Today Review the basics of an experiment Learn how to create a unit-weighted composite variable and how/why it is used in psychology. Learn how.
Multiple linear indicators A better scenario, but one that is more challenging to use, is to work with multiple linear indicators. Example: Attraction.
Creating Assessments AKA how to write a test. Creating Assessments All good assessments have three key features: All good assessments have three key features:
Teacher Engagement Survey Results and Analysis June 2011.
Chapter 4 – Research Methods in Clinical Psych Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Multiple linear indicators A better scenario, but one that is more challenging to use, is to work with multiple linear indicators. Example: Attraction.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Reliability, the Properties of Random Errors, and Composite Scores Week 7, Psych R. Chris Fraley
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Chapter 2: Behavioral Variability and Research Variability and Research 1. Behavioral science involves the study of variability in behavior how and why.
Chapter 10: Measurement Slavin Reliability & Validity Types of Measures Behavioral Observation Systems Sample Size Determination.
Psychology 3051 Psychology 305: Theories of Personality Lecture 2.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Measurement Theory in Marketing Research. Measurement What is measurement?  Assignment of numerals to objects to represent quantities of attributes Don’t.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Discussion Overview: Measurement I) Reliability of Measures I) Reliability of Measures II) Construct Validity II) Construct Validity III) Measurement scales.
Psychology 3051 Psychology 305: Theories of Personality Lecture 2.
Reliability: Introduction. Reliability Session Definitions & Basic Concepts of Reliability Theoretical Approaches Empirical Assessments of Reliability.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Psychology 3051 Psychology 305A: Theories of Personality Lecture 2 1.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
Reliability. Basics of test score theory Each person has a true score that would be obtained if there were no errors in measurement. However, measuring.
Measurement and Scaling Concepts
VALIDITY What is validity? What are the types of validity? How do you assess validity? How do you improve validity?
Reliability Analysis.
Ch. 5 Measurement Concepts.
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
پرسشنامه کارگاه.
5. Reliability and Validity
Questionnaire Reliability
RESEARCH METHODS Lecture 18
Personality An individual’s characteristic pattern of thinking, feeling, and acting.
Reliability Analysis.
Reliability, the Properties of Random Errors, and Composite Scores
Psychological Measurement: Reliability and the Properties of Random Errors The last two lectures were concerned with some basics of psychological measurement:
Measurement Concepts and scale evaluation
Presentation transcript:

Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability (i.e., the degree to which measurements are free of random error).

Why reliability alone is not enough Understanding the degree to which measurements are reliable, however, is not sufficient for evaluating their quality. In-class scale example –Recall that test-retest estimates of reliability tend to range between 0 (low reliability) and 1 (high reliability) –Note: The on-line correlation calculator is available at

Validity In this example, the measurements appear reliable, but there is a problem... Validity reflects the degree to which measurements are free of both random error, E, and systematic error, S. O = T + E + S Systematic errors reflect the influence of any non-random factor beyond what we’re attempting to measure.

Validity: Does systematic error accumulate? Question: If we sum or average multiple observations (i.e., using a multiple indicators approach), how will systematic errors influence our estimates of the “true” score?

Validity: Does error accumulate? Answer: Unlike random errors, systematic errors accumulate. Systematic errors exert a constant source of influence on measurements. We will always overestimate (or underestimate) T if systematic error is present.

Note: Each measurement is 2 points higher than the true value of 10. The errors do no average out.

Note: Even when random error is present, E averages to 0 but S does not. Thus, we have reliable measures that have validity problems.

Validity: Ensuring validity What can we do to minimize the impact of systematic errors? One way to minimize their impact is to use a variety of indicators Different kinds of indicators of a latent variable may not share the same systematic errors If true, then S will behave like random error across measurements (but not within measurements)

Example As an example, let’s consider the measurement of self-esteem. –Some methods, such as self-report questionnaires, may lead people to over-estimate their self-esteem. Most people want to think highly of themselves. –Other methods, such as clinical ratings by trained observers, may lead to under-estimates of self-esteem. Clinicians, for example, may be prone to assume that people are not as well-off as they say they are.

Note: Method 1 systematically overestimates T whereas Method 2 systematically underestimates T. In combination, however, those systematic errors cancel out. Clinical ratings Self- reports

Another example One problem with the use of self-report questionnaire rating scales is that some people tend to give high (or low) answers consistently (i.e., regardless of the question being asked).

ItemTSO I think I am a worthwhile person I have high self-esteem.4+15 I am confident in my ability to meet challenges in life My friends and family value me as a person Average score:4+15 In this example, we have someone with relatively high self- esteem, but this person systematically rates questions one point higher than he or she should. 1 = strongly disagree | 5 = strongly agree

ItemTSO I think I am a worthwhile person I have high self-esteem.4+15 I am NOT confident in my ability to meet challenges in life My friends and family DO NOT value me as a person Average score:4+14 If we “reverse key” half of the items, the bias averages out. Responses to reverse keyed items are counted in the opposite direction. T: ( [6-2] + [6-2]) / 4 = 4 O: ( [6-3] + [6-3]) / 4 = 4 1 = strongly disagree | 5 = strongly agree

Validity To the extent to which a measure has validity, we say that it measures what it is supposed to measure Question: How do you assess validity? ** Very tough question to answer! ** (But, we’ll give it a shot in our next class.)