A Method for Estimating the Correlations Between Observed and IRT Latent Variables or Between Pairs of IRT Latent Variables Alan Nicewander Pacific Metrics.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Lesson 10: Linear Regression and Correlation
Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
Logistic Regression Psy 524 Ainsworth.
Chapter 12 Simple Linear Regression
Statistical Techniques I EXST7005 Simple Linear Regression.
IRT Equating Kolen & Brennan, IRT If data used fit the assumptions of the IRT model and good parameter estimates are obtained, we can estimate person.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Correlation and regression Dr. Ghada Abo-Zaid
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
P M V Subbarao Professor Mechanical Engineering Department
Cox Model With Intermitten and Error-Prone Covariate Observation Yury Gubman PhD thesis in Statistics Supervisors: Prof. David Zucker, Prof. Orly Manor.
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
Models with Discrete Dependent Variables
The Simple Linear Regression Model: Specification and Estimation
Chapter 4 Multiple Regression.
Comparison of Reliability Measures under Factor Analysis and Item Response Theory —Ying Cheng , Ke-Hai Yuan , and Cheng Liu Presented by Zhu Jinxin.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
1 A MONTE CARLO EXPERIMENT In the previous slideshow, we saw that the error term is responsible for the variations of b 2 around its fixed component 
Discriminant Analysis Testing latent variables as predictors of groups.
Correlation Question 1 This question asks you to use the Pearson correlation coefficient to measure the association between [educ4] and [empstat]. However,
Ordinary Least Squares
Calibration & Curve Fitting
Correlation and Linear Regression
Equations in Simple Regression Analysis. The Variance.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
3.1 Ch. 3 Simple Linear Regression 1.To estimate relationships among economic variables, such as y = f(x) or c = f(i) 2.To test hypotheses about these.
Chapter 13: Inference in Regression
Chapter 11 Simple Regression
Introduction to plausible values National Research Coordinators Meeting Madrid, February 2010.
CORRELATION & REGRESSION
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Chapter 13 Statistics © 2008 Pearson Addison-Wesley. All rights reserved.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 9 Section 1 – Slide 1 of 39 Chapter 9 Section 1 The Logic in Constructing Confidence Intervals.
Montecarlo Simulation LAB NOV ECON Montecarlo Simulations Monte Carlo simulation is a method of analysis based on artificially recreating.
Applications The General Linear Model. Transformations.
1 G Lect 8b G Lecture 8b Correlation: quantifying linear association between random variables Example: Okazaki’s inferences from a survey.
1 Exploratory & Confirmatory Factor Analysis Alan C. Acock OSU Summer Institute, 2009.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Example x y We wish to check for a non zero correlation.
Joint Moments and Joint Characteristic Functions.
Measurement Math DeShon Univariate Descriptives Mean Mean Variance, standard deviation Variance, standard deviation Skew & Kurtosis Skew & Kurtosis.
Principal Component Analysis
Lecture 39 Numerical Analysis. Chapter 7 Ordinary Differential Equations.
Two Approaches to Estimation of Classification Accuracy Rate Under Item Response Theory Quinn N. Lathrop and Ying Cheng Assistant Professor Ph.D., University.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
Analysis of financial data Anders Lundquist Spring 2010.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
CFA with Categorical Outcomes Psych DeShon.
Data Modeling Patrice Koehl Department of Biological Sciences
The simple linear regression model and parameter estimation
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Part 5 - Chapter 17.
Evaluation of measuring tools: validity
Chapter 5 STATISTICS (PART 4).
Lecture Slides Elementary Statistics Thirteenth Edition
Statistical Methods For Engineers
Part 5 - Chapter 17.
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Ass. Prof. Dr. Mogeeb Mosleh
The Simple Linear Regression Model: Specification and Estimation
6.5 Taylor Series Linearization
5.2 Least-Squares Fit to a Straight Line
Correlation and Regression
Lecture # 2 MATHEMATICAL STATISTICS
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

A Method for Estimating the Correlations Between Observed and IRT Latent Variables or Between Pairs of IRT Latent Variables Alan Nicewander Pacific Metrics Presented at a conference to honor Dr. Michael W. Browne of the Ohio State University, September 9-10, 2010

Using the factor analytic version of item response (IRT) models, – Estimates of the correlations between the latent variables measured by test items are derived. – Also, estimates of the correlations between the latent variables measured by test items and external, observed variables are derived

Brief Derivations of the Correlations The normal ogive model for multiple-choice, dichotomous items may be written as, where, θ is the latent proficiency variable, a i is the item slope parameter, b i is the item location parameter, c i is a guessing parameter, and φ(t) is the normal density function.

Another useful version of this model is the so- called factor analytic representation: Let Y i be a latent response variable that is a linear function of θ plus error, where λ i may be considered as a factor loading and ε i is an error variable. It is further assumed that Y i and θ are normally distributed with zero means and unit variances, and that ε i is uncorrelated with θ and Y i.

Let γ i be a response threshold, defined so that if Yi > γ i the item is gotten correct and then (1) may be rewritten as, A graphical representation of this equation is given on the following slide. Then if λ i and γ i are rescaled as, And (1) becomes,

A Graph Showing Y i, the Latent Response Variable, Mapped into (1, 0) Using the Response Threshold, γ i

Estimating correlations between the latent variables measured by dichotomous items Suppose one wants to determine the correlation between the latent variables, θ i and θ j, that underlie the observed item responses, u i and u j. Let Y i and Y j be the latent response variables for the two MC items. or,

The resulting equation, does not seem useful in that it involves two, latent correlations, and However, from the definition of the tetrachoric correlation it follows that, where is the guessing-corrected tetrachoric correlation coefficient.

Estimating the correlations between the latent variables for dichotomous items and external, observed variables It is fairly easy to extend the logic above in order to derive a means for computing correlations between observed variables and IRT latent variables. First, define Z k as an observed variable scaled to have zero mean and unit variance. Then the correlation between Z k and the latent response variable, Y i, assumed to underlying a MC item, u i is given by,

Repeating the previous equation, we once again have an equation with two, latent correlation coefficients; However, following from the definition of a biserial correlation, we may substitute and obtain a solution involving only observables, where ρ* bis (Z k,u i ) is the guessing-corrected biserial correlation between the observed variables, Z k and u i.

Extending the latent-variable correlations to polytomous test items. In order to simplify exposition, only polytomous items having three categories are modeled. Generalization of the methods described below to items with more than three categories is very straightforward.

Let x ij be the score for item i scored in category j (j = 1, 2, …, m). Under commonly-used scoring rules, a three-category item would be scored 0, 1, or 2. As was done above, for the case of MC items having binary scores, let Y* i be the latent response variable underlying the polytomous item x ij : where λ i is a factor loading.

Let γ i1 and γ i2 be two response thresholds, defined so that: λ i and these two thresholds may be rescaled into IRT slope and location parameters, viz.

Fitting the previous model to data may be done with the nomal ogive version of Samejima’s (1969) Graded Response model, or the more commonly used logistic approximation thereof. However the correlations we are seeking here do not depend on the locations, but only the estimates of the, the item slope parameters.

First, define the correlation between the two latent response variables that underlie two, polytomous items, x ij and x kj, using previous logic, Then, solve for ρ(θ i θ k ) and substitute for λ i and λ k

Computing the correlation between an external variable and the latent variable measured by a polytomous item. From earlier derivations, it is fairly obvious that the correlation between an external variable Z k and the latent variable measured by a polytomous item, x ij is, where, ρ poly_s (Z k, x ij ) is the polyserial correlation between the external variable and the score on the polytomous item.

Some Numerical Examples Computing the correlations between the latent variables measured by three polytomous items, each having three categories. Ten replications of 300 observations were simulated using the values of a, b 1 and b 2 given below in Table 1, and with true values of ρ(θ 1,θ 2 ) = ρ(θ 2,θ 3 ) =.6, and ρ(θ 1,θ 3 ) = 1.

θ1θ1 θ2θ2 θ3θ3 θ4θ4 θ5θ5 Z1Z1.4 Z2Z2.2 Z3Z Z4Z4.4 Z5Z5.2 Z6Z Summary of Regression Slopes of Five Latent Variables, θ i, on Six Observed Variables, Z k

The coefficients developed in this inquiry have a very simple form for two basic reasons: 1.They make strong assumptions about data and, 2.The exotic correlation coefficients on which they are based (tetrachoric-polychoric and biserial- polybiserial) do the “heavy lifting”, mathematically, because of the complex calculations that are entailed in their computation. 3.It is also the case that the standard errors of these coefficients are rather large, and they almost certainly will require large samples for accuracy.