Construct Validity and Measurement

Slides:



Advertisements
Similar presentations
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Advertisements

Threats to Construct Validity
Validity (cont.)/Control RMS – October 7. Validity Experimental validity – the soundness of the experimental design – Not the same as measurement validity.
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Testing What You Teach: Eliminating the “Will this be on the final
Measurement Reliability and Validity
Using.  Purpose ▪ What question were the authors trying to answer? ▪ What was the primary relationship being studied?  Methods ▪ Subtitles: Participants;
Grant Weaver, Sam Vaninger, Brett Vihnanek.  This study was meant to examine the acute effects of a caffeine- containing supplement on upper and lower.
Chapter 8 Construct and External Validity in Experimental Research ♣ ♣ Construct Validity   External Validity   Cautions in Evaluating the External.
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Other Measurement Validity Types. OverviewOverview l Face validity l Content validity l Criterion-related validity l Predictive validity l Concurrent.
Correlation AND EXPERIMENTAL DESIGN
RESEARCH METHODS Lecture 18
VALIDITY.
Validity Validity of measurement Validity of research
A bunch of stuff you need to know
SOWK 6003 Social Work Research Week 4 Research process, variables, hypothesis, and research designs By Dr. Paul Wong.
Validity of Selection. Objectives Define Validity Relation between Reliability and Validity Types of Validity Strategies.
Chapter 7 Evaluating What a Test Really Measures
Reliability and Validity. Criteria of Measurement Quality How do we judge the relative success (or failure) in measuring various concepts? How do we judge.
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Construct Validity By Michael Kotutwa Johnson Submitted October 23, 2006 AED 615 Professor Franklin.
Understanding Validity for Teachers
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
1 Evaluating Psychological Tests. 2 Psychological testing Suffers a credibility problem within the eyes of general public Two main problems –Tests used.
Measurement in Exercise and Sport Psychology Research EPHE 348.
Ch 6 Validity of Instrument
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Technical Adequacy Session One Part Three.
Methodology Experiments.
Final Study Guide Research Design. Experimental Research.
Validity & Practicality
 Internal Validity  Construct Validity  External Validity * In the context of a research study, i.e., not measurement validity.
Today: Our process Assignment 3 Q&A Concept of Control Reading: Framework for Hybrid Experiments Sampling If time, get a start on True Experiments: Single-Factor.
The Basics of Experimentation Ch7 – Reliability and Validity.
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Validity Is the Test Appropriate, Useful, and Meaningful?
EDU 8603 Day 6. What do the following numbers mean?
Assessing the Quality of Research
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Validity RMS – May 28, Measurement Reliability The extent to which a measurement gives results that are consistent.
Measurement Validity.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Reading and Evaluating Research Method. Essential question to ask about the Method: “Is the operationalization of the hypothesis valid? Sections: Section.
KNR 295 Measurement Slide 1 Measurement Theory & Construct Validity Chapter 3.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Chapter 6 - Standardized Measurement and Assessment
MEASUREMENT: PART 2. Overview  Measurement Validity  Face Validity (non-statistical)  Content Validity (mostly non-statistical)  Construct Validity.
Experiments.  Labs (update and questions)  STATA Introduction  Intro to Experiments and Experimental Design 2.
Validity & Reliability. OBJECTIVES Define validity and reliability Understand the purpose for needing valid and reliable measures Know the most utilized.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Construct validity s.net/kb/consthre.htm.
Reliability and Validity
Measurement: Part 2.
Measurement: Part 2.
Journalism 614: Reliability and Validity
Introduction to Measurement
پرسشنامه کارگاه.
RESEARCH METHODS Lecture 18
Experiments: Validity, Reliability and Other Design Considerations
Experiments II: Validity and Design Considerations
Measurement: Part 2.
Cal State Northridge Psy 427 Andrew Ainsworth PhD
Presentation transcript:

Construct Validity and Measurement Do they measure what they say they measure?

Construct Validity Cause Construct Effect Construct operationalization What you think What you think Cause Construct Effect Construct Theory operationalization Program Observations Observation What you do What you see

Can we generalize back to the constructs? Construct Validity What you think What you think Cause Construct Effect Construct Theory Can we generalize back to the constructs? Program Observations Observation What you do What you see

Construct Validity the degree to which inferences can legitimately be made from the operationalizations in a study to the theoretical constructs on which they are based or Did the study really measure what it claimed to measure?

Central Questions “Is your operationalization an accurate translation of the construct?” “Are you measuring what you intended to measure?” “Does your program/treatment accurately reflect what you intended?”

Construct Validity translation validity criterion-related validity face validity content validity criterion-related validity predictive validity concurrent validity convergent validity discriminant validity Important Point all aspects or elements of construct validity think of these as things to look for when evaluating construct validity

Translation Validity vs. Criterion-Related Validity assesses whether the operationalization is a good reflection of the construct criterion validity… assesses whether the operationalization behaves in the way it should given your theory of the construct (uses other measures (criteria) to assess construct validity)

Translation Validity Focuses on whether the operationalization is a good reflection of the construct Face Validity: On its face, does the operationalization look like a good translation of the construct?

Translation Validity: Face validity Construct satisfaction with KNR 164 course Possible Operationalizations (ways to measure construct) Question: Which of these are more or less reasonable on the face of it?

Translation Validity: Face validity Construct Fitness Possible Operationalizations (ways to measure construct) Question: Which of these are more or less reasonable on the face of it?

Translation Validity: Content Validity Content Validity: Operationalization is checked against the relevant content domain for the construct Often involves researching just how the construct is defined by those in a position to know (experts)

Translation Validity: Content Validity Construct fitness program Possible Operationalization (key elements?) Question: Are these the correct elements of the construct?

Criterion Validity The performance of your operationalization (i.e., measure) is checked against a criterion. A prediction of how the operationalization will perform on some other measure based on your theory or construct “Validating” a measure based on its relationship to another measure

Criterion Validity Predictive Validity: Operationalization’s ability to predict something it should theoretically be able to predict e.g.: GRE and grad school performance (!) Concurrent Validity: Operationalization’s ability to distinguish between groups which theoretically should be different e.g.: fatness test for athletes and non-athletes

Criterion Validity Convergent Validity: Degree to which the operationalization is similar to (converges on) other operationalizations to which it theoretically should be similar e.g.: Discriminant Validity: Degree to which the operationalization is not similar to other operationalizations to which it theoretically should not be similar

Threats to Construct Validity inadequate preoperational explication mono-operation bias mono-method bias interaction of different treatments interaction of testing and treatment restricted generalizability across constructs confounding constructs and levels social threats to construct validity

Examples of Threats to Construct Validity construct not defined clearly enough only one possible example of the construct (either IV or DV)

Examples of Threats to Construct Validity inaccurate labeling of construct missing important elements failure to define or consider “dose” social issues participants guessing what they are “supposed” to do or say participants being apprehension experimenter’s expectancies biasing observations being made

Below is a research problem Below is a research problem. Identify which of the threats to construct validity may be of major concern. General idea behind the research scenario (a quotation from our researcher): “I feel that plyometric strength training is more effective for gaining strength than isometric strength training. I’ve done plyometrics for years, and it has worked wonders.” An undergrad class taught by the researcher is split into 3 groups of 30. One third is assigned to a plyometric strength-training program, 1/3 to an isometric program, and 1/3 do nothing. Before assigning them, the researcher makes sure to tell the entire class about the purpose of the research, and explains we are doing it to see if the researcher’s suspicions about plyometrics are correct. Before and at the end of the programs, all students are tested on a measure of strength - a grip dynamometer. This test is supervised by the researcher to make sure proper procedures are followed. It is expected that the plyometric group will make the greater strength gains.

Guiding Questions Construct Validity –Measures/Observations What, in theory, are the researchers trying to assess or measure (list each construct)  Answer the following questions for each construct included in the study. Do the researchers explicitly define the construct? If so, how? If the construct is not explicitly defined by the authors, what does the construct mean to you (in theory)? How did the researchers operationalize the construct? That is, how exactly did they measure/assess the construct? In your opinion, is the operationalization of the construct a reasonable approximation of the theoretical construct? In other words, does the measure they used in the study match up with what they said they were trying to measures? [This is the key Construct Validity question] Do you see any limitation with their operationalization (i.e., are any of the common threats an issue)? Do the researchers provide or present any evidence of content validity or criterion-related validity?

Guiding Questions Construct Validity – Interventions/Treatments What, in theory, are the researchers trying to test as their intervention or treatment? Do the researchers provide some idea of what their intervention/treatment should look like in theory? If not, what should the intervention/treatment be to you (in theory)? How did the researchers operationalize the intervention/treatment? That is, what exactly did they have the subjects or groups of subjects do? In your opinion, is the operationalization of the intervention/treatment a reasonable approximation of what the research wanted to test in theory? In other words, does the intervention/treatment they used in the study match up with what they said they were trying to do? [This is the key Construct Validity question] Do you see any limitation with their operationalization (i.e., are any of the common threats an issue)?

Overall (final) question in each case: Do the limitations to construct validity: change the meaning of the study’s conclusions or have the potential to alter the results of the study.

Practice Activity #1: Evaluate the construct validity of this study General idea behind the research scenario (a quotation from our researcher): “I feel that plyometric strength training is more effective for gaining strength than isometric strength training. I’ve done plyometrics for years, and it has worked wonders.” An undergrad class taught by the researcher is split into 3 groups of 30. One third is assigned to a plyometric strength-training program, 1/3 to an isometric program, and 1/3 do nothing. Before assigning them, the researcher makes sure to tell the entire class about the purpose of the research, and explains we are doing it to see if the researcher’s suspicions about plyometrics are correct. Before and at the end of the programs, all students perform a 1RM leg extension test as a measure of strength. This test is supervised by the researcher to make sure proper procedures are followed. During the 4 weeks of training, the subjects in the plyometric group did 10 drop jumps each day from a height of 2 feet, while the subjects in the isometric group performed 3 sets of 10 reps of the following exercises using Nautilus equipment: bench press, shoulder press, and biceps curls. Those in the control group were instructed to do no physical activity during the 4 weeks. It was expected that the plyometric group will make the greatest strength gains.

Practice Activity #2 Use the guiding questions to evaluate the construct validity of measures used in the study distributed in class