Reliability & Validity the Bada & Bing of YOUR tailored survey design.

Slides:



Advertisements
Similar presentations
Chapter 8 Flashcards.
Advertisements

Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Topics: Quality of Measurements
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
Some (Simplified) Steps for Creating a Personality Questionnaire Generate an item pool Administer the items to a sample of people Assess the uni-dimensionality.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
VALIDITY AND RELIABILITY
Research Methodology Lecture No : 11 (Goodness Of Measures)
Reliability & Validity.  Limits all inferences that can be drawn from later tests  If reliable and valid scale, can have confidence in findings  If.
Measurement Reliability and Validity
Reliability, the Properties of Random Errors, and Composite Scores.
CH. 9 MEASUREMENT: SCALING, RELIABILITY, VALIDITY
Reliability and Validity of Research Instruments
RESEARCH METHODS Lecture 18
Measurement: Reliability and Validity For a measure to be useful, it must be both reliable and valid Reliable = consistent in producing the same results.
SOWK 6003 Social Work Research Week 5 Measurement By Dr. Paul Wong.
Validity, Reliability, & Sampling
Classroom Assessment A Practical Guide for Educators by Craig A
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Classroom Assessment Reliability. Classroom Assessment Reliability Reliability = Assessment Consistency. –Consistency within teachers across students.
Technical Issues Two concerns Validity Reliability
Validity and Reliability
Reliability, Validity, & Scaling
PhD Research Seminar Series: Reliability and Validity in Tests and Measures Dr. K. A. Korb University of Jos.
VALIDITY, RELIABILITY, and TRIANGULATED STRATEGIES
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Data Analysis. Quantitative data: Reliability & Validity Reliability: the degree of consistency with which it measures the attribute it is supposed to.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
MGTO 324 Recruitment and Selections Validity I (Construct Validity) Kin Fai Ellick Wong Ph.D. Department of Management of Organizations Hong Kong University.
Research Methodology Lecture No :24. Recap Lecture In the last lecture we discussed about: Frequencies Bar charts and pie charts Histogram Stem and leaf.
Tests and Measurements Intersession 2006.
Validity In our last class, we began to discuss some of the ways in which we can assess the quality of our measurements. We discussed the concept of reliability.
Measurement Validity.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Reliability, the Properties of Random Errors, and Composite Scores Week 7, Psych R. Chris Fraley
Class 9 Dependent Variables, Instructions/Literature Review
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
The Theory of Sampling and Measurement. Sampling First step in implementing any research design is to create a sample. First step in implementing any.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Reliability: The degree to which a measurement can be successfully repeated.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Measurement Issues General steps –Determine concept –Decide best way to measure –What indicators are available –Select intermediate, alternate or indirect.
Assessment: Reliability & Validity. Reliability Refers to the repeatability of a given testing instrument The extent to which a student would be expected.
Experimental Research Methods in Language Learning Chapter 12 Reliability and Reliability Analysis.
©2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Reliability: Introduction. Reliability Session 1.Definitions & Basic Concepts of Reliability 2.Theoretical Approaches 3.Empirical Assessments of Reliability.
Measurement Experiment - effect of IV on DV. Independent Variable (2 or more levels) MANIPULATED a) situational - features in the environment b) task.
Reliability a measure is reliable if it gives the same information every time it is used. reliability is assessed by a number – typically a correlation.
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Reliability EDUC 307. Reliability  How consistent is our measurement?  the reliability of assessments tells the consistency of observations.  Two or.
Measurement and Scaling Concepts
Class 9 Dependent Variables, Instructions/Literature Review Class 9 Dependent Variables, Instructions/Literature Review Chapters 13 Spring 2016.
MGMT 588 Research Methods for Business Studies
Reliability and Validity
Assessment Theory and Models Part II
Test Validity.
Journalism 614: Reliability and Validity
Reliability & Validity
Classroom Assessment Validity And Bias in Assessment.
Introduction to Measurement
Measurement Characteristics of Client Assessment
Reliability and Validity of Measurement
RESEARCH METHODS Lecture 18
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Reliability & Validity the Bada & Bing of YOUR tailored survey design

This presentation has been influenced  not at all  a little bit  CONSIDERABLY by the work & wisdom of Dan Koretz. Thanks! Pre-Presentation credit

Core concept of validity You wish to measure a construct, but can never know true score for sure (e.g. 6th grade math proficiency, self esteem) You must draw an inference about the construct based on a sample or indicator of behavior-- something you can actually “touch” (e.g., 6th grade math test, self esteem survey) VALIDITY describes how well performance on your indicator justifies your inference about the construct

Error: Validity’s arch nemesis Sampling error: occurs from sampling units of observation (i.e, populations of humans) Measurement error (M.E.): occurs across instances of measurement from each unit (i.e. individual humans)

More on measurement error Outcome i = True score + M.E. Measurement error can be… systematic (does not “wash out” across repeated measurements) or random (does “wash out” across repeated measurements)

Systematic error Chris repeatedly takes a self esteem survey written in Latin Outcome repeatedly affected by factor not relevant to construct being measured “Construct irrelevant variance” **Note: the only Latin that Chris speaks involves inesway**

Systematic error cont’d “Construct Under-representation” If construct is poorly represented, repeated measurements will not converge on true score with regard to entire construct e.g. global self esteem survey asks only about Chris’ confidence on golf course

Random error Chris repeatedly takes self esteem survey Sometimes mood = Sometimes mood = Over time, outcomes will converge on Chris’ true score

Some details Validity is an attribute of your inference, not the instrument itself, which may be more valid for some inferences and/or populations than others (e.g. self-esteem survey in Latin). Validity is not an all or nothing phenomenon, but a matter of degree, and we must piece together evidence suggesting how valid our inferences may or may not be.

Types of validity/validity evidence note: other terms exist, but this will be the focus of S-015

Types of validity/validity evidence Construct Validity: How well does performance on our instrument justify inferences about the construct? “Validity” What we’re ultimately shooting for

Types of validity/validity evidence Content based evidence (a/k/a content validation study): Compare your instrument to your very thoroughly defined construct… Does the instrument adequately represent the construct? Harder than it seems (constructs can be messy)

Types of validity/validity evidence Convergent-discriminant evidence: Measures of similar constructs should converge. Measures of less similar constructs should diverge. (e.g. Two math tests should correlate more strongly than a math and reading test)

Multitrait-Multimethod Matrix (MTMM) A fun way to display convergent- discriminant validity (or not) Pass out hand outs: Now

But alas, complications abound… What constitutes “similar” What constitutes “less similar” What constitutes “convergence” What constitutes “divergence” ????

Plausible toy correlations Math 1 Math Math 2.82 Read 1.74 Read 2.70

Closing thoughts on validity We must piece together evidence that is often murky and incomplete to reach judgment. An instrument that is fairly valid for one use, inference, or population may not be valid for others.

Oh, BTW… The more reliable your instrument, the better your chance of drawing fairly valid inferences. (Old Faithful) 

Core concept of reliability Reliability is consistency of results across repeated measurements (e.g. assuming no interventions or natural attitudinal shifts in between, a subject taking a highly reliable survey would perform quite similarly each time s/he took it.)

Some details Reliability is also a matter of degree, often expressed as a coefficient ranging from A test or survey may be more reliable for some populations than others (e.g. surveys tend to be more reliable among older/more educated populations.)

POP QUIZ True or false… 1.) An instrument that allows us to draw a reasonably valid inference must be reasonably reliable 2.) A reasonably reliable instrument must allow us to draw a reasonably valid inference

POP QUIZ cont’d Regarding R & V, how might one describe…

A few (of many) ways to assess reliability Assess internal consistency Assuming a survey taps one and only one construct, the results from the first half should correlate highly with results from the second half; the odd items should correlate highly with the evens, etc. (split-half correlations)

Coefficient Alpha a/k/a Cronbach’s alpha The average of all possible split-half correlations in a given sample generally preferred to single split-half correlation

A few (of many) ways to assess reliability Test-retest Assuming no interventions or natural shifts in attitude, a reasonably reliable survey will yield similar results from the same person across repeated administrations

WARNING A test or survey with a high reliability coefficient does not guarantee that your results will be highly reliable. (e.g. Differences in administrative conditions can effect the consistency of your results across repeated administrations.

Questions???

References, etc Linn, R.L. & Gronlund, N. E. (2000). Measurement and Assessment in Teaching, 8 th Ed. New Jersey: Prentice-Hall, Inc. See diagram displayed on page 75 of the reference textbook. A very cool link that covers a TON of stuff on all forms of social research…