RCAL1 Validity Threats, Biases & Systematic Error in International Research Roger Calantone 2004 Not Complete Without Commentary.

Slides:



Advertisements
Similar presentations
Scale Construction and Halo Effect in Secondary Student Ratings of Teacher Performance Ph.D. Dissertation Defense Eric Paul Rogers Department of Instructional.
Advertisements

Agenda Levels of measurement Measurement reliability Measurement validity Some examples Need for Cognition Horn-honking.
Cross Cultural Research
Conceptualization and Measurement
The Research Consumer Evaluates Measurement Reliability and Validity
Taking Stock Of Measurement. Basics Of Measurement Measurement: Assignment of number to objects or events according to specific rules. Conceptual variables:
Research Methodology Lecture No : 11 (Goodness Of Measures)
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Experimental Research Designs
Measurement. Scales of Measurement Stanley S. Stevens’ Five Criteria for Four Scales Nominal Scales –1. numbers are assigned to objects according to rules.
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
Culture and psychological knowledge: A Recap
Assessment: Reliability, Validity, and Absence of bias
Chapter 4 Validity.
Test Validity: What it is, and why we care.
Validity, Sampling & Experimental Control Psych 231: Research Methods in Psychology.
Validity, Sampling & Experimental Control Psych 231: Research Methods in Psychology.
Beginning the Research Design
MSc Applied Psychology PYM403 Research Methods Validity and Reliability in Research.
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Introduction to Communication Research
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Chapter 9 Experimental Research Gay, Mills, and Airasian
CAUSAL-COMPARATIVE RESEARCH Prepared for: Eddy Luaran Prepared by: Nur Hazwani Mohd Nor ( ) Noriziati Abd Halim ( ) Noor fadzilah.
VALIDITY. Validity is an important characteristic of a scientific instrument. The term validity denotes the scientific utility of a measuring instrument,
EVALUATING YOUR RESEARCH DESIGN EDRS 5305 EDUCATIONAL RESEARCH & STATISTICS.
Chapter 8 Experimental Research
© 2013 Cengage Learning. Outline  Types of Cross-Cultural Research  Method validation studies  Indigenous cultural studies  Cross-cultural comparisons.
Measurement in Exercise and Sport Psychology Research EPHE 348.
Translation and Cross-Cultural Equivalence of Health Measures.
LECTURE 06B BEGINS HERE THIS IS WHERE MATERIAL FOR EXAM 3 BEGINS.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Cross-Cultural Research Methods. Methodological concerns with Cross-cultural comparisons  Equivalence  Response Bias  Interpreting and Analyzing Data.
ScWk 240 Week 6 Measurement Error Introduction to Survey Development “England and America are two countries divided by a common language.” George Bernard.
Validity Is the Test Appropriate, Useful, and Meaningful?
SURVEY RESEARCH.  Purposes and general principles Survey research as a general approach for collecting descriptive data Surveys as data collection methods.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Chapter 8 Marketing Research and Information System.
Measurement Validity.
Research in Communicative Disorders1 Research Design & Measurement Considerations (chap 3) Group Research Design Single Subject Design External Validity.
Learning Objective Chapter 9 The Concept of Measurement and Attitude Scales Copyright © 2000 South-Western College Publishing Co. CHAPTER nine The Concept.
AMSc Research Methods Research approach IV: Experimental [1] Jane Reid
PSYC313-04B Cross-Cultural Studies The Emic-Etic Dilemma: Measurement problems in cross- cultural research Michael Hills, PhD.
Personality across cultures Ype H. Poortinga (Prof Em) Tilburg University, Netherlands & University of Leuven, Belgium.
META-ANALYSIS, RESEARCH SYNTHESES AND SYSTEMATIC REVIEWS © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON.
Scientific Methods and Terminology. Scientific methods are The most reliable means to ensure that experiments produce reliable information in response.
Question paper 1997.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Measurement and Scaling
Chapter 10 Experimental Research Gay, Mills, and Airasian 10th Edition
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
Ch 9 Internal and External Validity. Validity  The quality of the instruments used in the research study  Will the reader believe what they are readying.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Translation and Cross-Cultural Equivalence of Health Measures
GENERALIZING RESULTS: the role of external validity.
CHAPTER 8 EXPERIMENTS.
8 Experimental Research Design.
Concept of Test Validity
Experimental Research Designs
Reliability & Validity
Introduction to Measurement
Chapter Eight: Quantitative Methods
Reliability and Validity of Measurement
Chapter 6 Research Validity.
Experimental Research
Experimental Research
Presentation transcript:

RCAL1 Validity Threats, Biases & Systematic Error in International Research Roger Calantone 2004 Not Complete Without Commentary

RCAL2 Overarching Empirical Goal Is Cross-cultural Equivalence In: Measures & their properties Instruments & their administration Researcher / experimenter conduct Research settings

RCAL3 We Must Be Able to Set up Controls to Positively Insure Equivalence. Failing This We Must Have Specific Measures of These Threats to Validity, So We Might Control for Them Statistically.

RCAL4 Bias Cross-cultural bias is the asymmetry of measurement across cultures. Bias can be a function of the instrument based constructs interacting differentially with the cultures addressed, the instrument itself or both. A lack of cross-cultural bias permits measurement equivalence, but cannot guarantee it.

RCAL5 Equivalence Construct equivalence (also referred to as structural equivalence). Measurement unit equivalence. Scalar equivalence (also referred to as full score comparability). The above levels of equivalence are listed in increasing order of difficulty.

RCAL6 Construct Equivalence The most basic level of equivalence. The same construct occurs in each culture, although it may be measured in a different manner (items or methods). (Of course you may not be able to compare scores across cultures unless the constructs are operationalized the same way.).

RCAL7 Measurement Unit Equivalence Constructs are sufficiently similar and well used enough across the cultures so offset mechanism is known or discoverable. Simple example is Celsius versus Kelvin scales of temperature.(Simple transform universally). Complex example is use of a Swedish IQ test on Turkish youngsters.(A constant disadvantage is known but varies by sample and/or by study).

RCAL8 Scalar Equivalence Hardest to obtain; Rarely demonstrated. Measurement instrument is on the same ratio scale in each cultural group. (eg:measure of body length & weight). Achievable when scores on an instrument have same interval scale across cultural groups. (Refers to useful characteristic of interval scales => differences on an interval scale are measured at a ratio level).

RCAL9 Equivalence Must Be Established It is not an intrinsic property of a scale or instrument or method. There is no evidence that equivalence in one set of cultural comparisons is present across other cultural comparisons. Equivalence is a function of the properties of an instrument and the characteristics of the cultural groups involved in the study.

RCAL10 Claims of Equivalence Can Be Controversial ! Scalar equivalence is often claimed when only construct equivalence has been established. Many times this is based on finding similar item loadings of similar or even identical items, using an EFA on the several cultural groups. This is very open to suspicion.

RCAL11 2 Big Problems Using EFA to Establish Scalar Equivalence Measurement units may not be the same across groups. If the scores in one group are doubled, but the correlation pattern remains the same, the cross-cultural differences are affected but the loadings will be unaffected ! Measurement unit equiv.. May obtain, but a bias that affects all of an instrument’s stimuli in a systematic way cannot be detected by EFA.

RCAL12 EFA Problems Continued Both arguments against the use of EFA to establish scalar equivalence derive from the use of the correlation matrix as input to an EFA. Linear transformations, e.g.: (y=ax+b with a>0) impact raw scores & hence x-cultural differences but do not impact correlations. Techniques insensitive to such linear transformations are ill-suited to substantiate scalar equivalence.

RCAL13 Avoiding EFA The simple fix is to design for scalar equivalence from the beginning and to only attempt scalar testing using CFA, which uses the covariance matrix and permits statistical equivalence tests of loadings across cultural groups. Unit measure equivalence testing should also employ Optimal Scaling.

RCAL14 Biases & Their Detection Bias is the basic name for all the factors that challenge the validity of cross- cultural comparisons. Three types of Bias are commonly recognized: Construct, Method, & Item. Bias is always present, but with care, can almost always be mitigated to non- threatening levels through good design.

RCAL15 TABLE 1 Overview of Types of Bias and Their Most Common Causes (after van de Vijver & Poortinga, 1997, p. 26) Type of BiasSource Construct incomplete overlap of definitions of the construct across cultures differential appropriateness of (sub)test content (e.g., skills do not belong to the repertoire of one of the cultural groups) poor sampling of all relevant behaviors (e.g., short instruments) incomplete coverage of the construct (e.g., not all relevant domains are sampled) Method differential social desirability differential response styles such as extremity scoring and acquiescence differential stimulus familiarity lack of comparability of samples (e.g.differ in educational background, age, or gender composition) differences in physical conditions of administration differential familiarity with response procedures tester/interviewer effects communication problems between respondent and interviewer in either cultural group Item poor item translation inadequate item formulation (e.g., complex wording) item(s) may invoke additional traits or abilities incidental differences in appropriateness of the item content (e.g., topic of item of educational test not in curriculum in one cultural group)

RCAL16 TABLE 2 Is Level of Equivalence Affected by Bias? Level of Equivalence Construct Measurement Unit a Scalar a, b Type of Bias Construct biasno no no Method bias: uniformyes yes no nonuniformyes no no Item bias: uniformyes yes no nonuniformyes no no a. The same measurement unit is assumed in each cultural group. b. The same origin is assumed in each cultural group.

RCAL17 Construct Bias Occurs when construct measured is not identical across groups. (appears frequently in comparisons of western & non-western cultures). Can be induced by lack of overlap in behaviors associated with the cultures studied. (E.g.. Machiavellianism). Also, Construct underrepresentation.

RCAL18 Construct Bias Fixes Emic approaches must precede. Decentered approach in conceptualization of the study. Convergence approach in conceptualization of the study. This is considerably more difficult if indigenous materials must be developed.

RCAL19 Method Bias Bias arises from characteristics of instrument and/or its administration. Differential response styles across cultures. (use of scales/extremes). Differential familiarity with stimuli used. Differential familiarity with response tool: E.g. draw a picture versus make a wire model.

RCAL20 Method Bias Avoidance “Culture-Free” tests. “Culture-Fair” tests. Interviewer repetition & rotation. Systematic variation of stimuli across cultures & examination of score changes as well as differential learning effects. Use of un-timed exercises.

RCAL21 Item Bias Differential appropriateness of item content (asymmetric item sets). Inadequate item functioning (e.G. Complex wording). Inadequate translation concordance. All of these can be mitigated by rigorous pretesting, with post-inquiry debriefing.

RCAL22 TABLE 3 Four Common Types of Cross-Cultural Studies Orientation More on Consideration of Contextual Factors Hypothesis TestingExploration No Generalizability Psychological differences Yes Theory-driven External validation SOURCE: van de Vijver & Leung,

RCAL23 Discussion Questions Can we really design culture-equivalent instruments? Can we reduce the bias in our comparative studies? Are instrument fixes better than statistical fixes? Where does equivalence and bias fit in the EMIC/ETIC debate?

RCAL24