Multidimensional Adaptive Testing with Optimal Design Criteria for Item Selection Joris Mulder & Wim J. Van Der Linden 1.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Statistical Decision Theory Abraham Wald ( ) Wald’s test Rigorous proof of the consistency of MLE “Note on the consistency of the maximum likelihood.
Component Analysis (Review)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Managerial Decision Modeling with Spreadsheets
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: asymptotic properties of estimators: plims and consistency Original.
« هو اللطیف » By : Atefe Malek. khatabi Spring 90.
3 rd Place Winning Project, 2009 USPROC Author: Kinjal Basu Sujayam Saha Sponsor Professor: S. Ghosh A.K. Ghosh Indian Statistical Institute, Kolkata,
Self Organization: Competitive Learning
Model Assessment, Selection and Averaging
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Visual Recognition Tutorial
Chapter 6: Multilayer Neural Networks
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Separate multivariate observations
A comparison of exposure control procedures in CATs using the 3PL model.
Normalised Least Mean-Square Adaptive Filtering
Photo-realistic Rendering and Global Illumination in Computer Graphics Spring 2012 Stochastic Radiosity K. H. Ko School of Mechatronics Gwangju Institute.
Summarized by Soo-Jin Kim
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Principles of Pattern Recognition
STOCHASTIC DOMINANCE APPROACH TO PORTFOLIO OPTIMIZATION Nesrin Alptekin Anadolu University, TURKEY.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 7 Sampling Distributions.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 6 Sampling Distributions.
Empirical Financial Economics Asset pricing and Mean Variance Efficiency.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Rasch trees: A new method for detecting differential item functioning in the Rasch model Carolin Strobl Julia Kopf Achim Zeileis.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 Advances in the Construction of Efficient Stated Choice Experimental Designs John Rose 1 Michiel Bliemer 1,2 1 The University of Sydney, Australia 2.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Calibration of Response Data Using MIRT Models with Simple and Mixed Structures Jinming Zhang Jinming Zhang University of Illinois at Urbana-Champaign.
Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 7 Sampling Distributions.
Simple regression model: Y =  1 +  2 X + u 1 We have seen that the regression coefficients b 1 and b 2 are random variables. They provide point estimates.
Discriminative Training and Acoustic Modeling for Automatic Speech Recognition - Chap. 4 Discriminative Training Wolfgang Macherey Von der Fakult¨at f¨ur.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
1 E. Fatemizadeh Statistical Pattern Recognition.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
1 MODELING MATTER AT NANOSCALES 4. Introduction to quantum treatments The variational method.
International Conference on Design of Experiments and Its Applications July 9-13, 2006, Tianjin, P.R. China Sung Hyun Park, Hyuk Joo Kim and Jae-Il.
BCS547 Neural Decoding.
Lecture 4 Linear machine
CpSc 881: Machine Learning
Chapter 13 (Prototype Methods and Nearest-Neighbors )
NATIONAL CONFERENCE ON STUDENT ASSESSMENT JUNE 22, 2011 ORLANDO, FL.
Autoregressive (AR) Spectral Estimation
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
IMPACT EVALUATION PBAF 526 Class 5, October 31, 2011.
Measurement and Scaling Concepts
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Bayesian Semi-Parametric Multiple Shrinkage
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Further Inference in the Multiple Regression Model
Chapter 2 Minimum Variance Unbiased estimation
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Learning From Observed Data
Presentation transcript:

Multidimensional Adaptive Testing with Optimal Design Criteria for Item Selection Joris Mulder & Wim J. Van Der Linden 1

The choice of criterion (D-optimality, A- optimality,…) should consider the goal of testing. A different optimal design criterion for item selection in MAT seems more appropriate. 2

Motivation To find the matches between the different cases of MCAT and the performance of optimal design criteria. To investigate the preference of the optimality criteria for items in the pool with specific patterns of parameter values: Will the criterion for selection in a MCAT program with nuisance abilities select only items that are informative about the intentional abilities? Are there any circumstances in which they also select items that are mainly sensitive to nuisance ability To report some features of Fisher information matrix and its use in adaptive testing that have been hardly noticed 3

Response Model 4

Fisher Information Each element in the matrix has a common factor: When selecting the kth item, the information is: on which different optimality criteria apply. 5

Item Information Matrix in MIRT The information matrix include: (1) Function of g, and (2) matrix aa’ Reparameterize into a one- dimensional function by substituting 6

Where is the Euclidean norm of a i. The ability value is determined by solving: The results: 7

Item selection criteria for MAT Three cases of multidimensional testing: (1) All abilities are intentional (2) Some ability are intentional and others are nuisance (3) All abilities are intentional, but the interest is only in a specific linear combination of them 8

All abilities Intentional D-Optimality (Segall, 1996) which can be expresses as The criterion tends to select items with a large discrimination parameter for the ability with a relatively large (asymptotic) variance for its current estimator (minimax mechanism) 9

Items with large discrimination parameters for more than one ability are generally not informative. Consequently, the criterion of D- optimality tends to prefer items that are sensitive to a single ability over items sensitive to multiple abilities (trade-off effect). Segall (1996) proposed a Bayesian version of D-optimality for MCAT. 10

A-Optimality: minimize the trace of the inverse of the information matrix This results contains the determinant of the information matrix as an important factor. And will similar to that of D-optimality. Can easily extend to a Bayesian version 11

E-Optimal: maximized the smallest eigenvalue of the information matrix. May behave unfavorably because the contribution of an item with equal discrimination parameters to the test information vanishes when the sampling variance of the ability estimator have become equal to each other. This fact contradicts the fundamental rule that the average sampling variance of ability should always decrease after a new observation. Using E-optimality for item selection in MCAT may result in occasionally bad item selection and its use not recommended. 12

Graphical Example 13

14 Item 1: a=(0.5,0) Col Item 2: a=(0.64,0.64) B&W

Nuisance Abilities 15

Both Ds-optimality and As-optimality generally selects items that highly discriminate with respect to the intentional ability. However, when the amount of information about the nuisance abilities is relatively low (that, determinant of nuisance ability is small), an item that highly discriminates with respect to the nuisance abilities is often preferred. 16

Composite ability C-optimality prefer items with discrimination parameters that reflect the weights of importance in the composite ability. Thus, items that with is generally more informative. 17

18 Labda=[1 1], a1=[0.5 1]’, a2=[0.8 1]’, labda*a1=1.5 > labda*a2=0.8. Labda=[1 1], a1=[0.5 1]’, a2=[ ]’, labda*a1=1.5 < labda*a2=1.6

Simulation Study Two dimensions MACT Item pool: 200 items generated from a1~N(1,0.3), a2~N(1,0.3), b~N(0,3) and 10c~Bin(3,0.5). Stopping rule: 30 items For each combination theta1 =-1,0,1 and theta2=- 1,0,1, a total of 100 adaptive test administration were simulated. Bias and MSE were compared between different criterion optimality Random selection was served as baseline. 19

20

Theta1 and theta2 intentional 21 A-optimality and D-optimality resulted more accurate ability estimation than E-optimality (which is even worse than R).

theta1 intentional and theta2 a nuisance Ds-optimality selects items that minimize the asymptotic variance of the intentional theta1 (the MSE of theta1 is smaller than that of theta1 when theta1 and theta 2 are both intentional). However, the MSE for the theta2 is much larger. 22

Composite ability 23 When equal weights, C-optimality with weights (1/2,1/2) yielded the highest accuracy for composite ability, however, larger MSE for separate abilities.

24 c(1,0) When unequal weights (3/4,1/4), the Ds-optimal was similar to c(3/4,1/4).

Average Values of Optimality Criteria 25 Except for E-optimality, each of the criterion produced the smallest average value for the specific quantity optimized by the criteria.

Conclusions When all abilities are intentional, both A-optimality and D-optimality result in the most accurate estimation for the separate abilities. The most informative items measure mainly one ability. Both criteria tend to “minimax”. When one of the abilities is intentional and the others are nuisance, item selection based on Ds-optimality (or As-optimality) result in the most accurate estimates for the intentional ability. Items that measure only the intentional ability are generally most informative. When the current inaccuracy of the estimator of a nuisance ability becomes too large relative to that of the intentional abilities, an item that sensitive to the nuisance ability will occasionally preferred. 26

For composition abilities, c-optimality with weights lambda proportional to the coefficients in the composite ability results in the most accurate estimation of ability. The criterion has a preference for items when the proportion of the discrimination parameters reflects the weights in the combination. Content control and exposure rate should be considered. CAT for ipsative tests 27