Mathematical Model for the Law of Comparative Judgment in Print Sample Evaluation Mai Zhou Dept. of Statistics, University of Kentucky Luke C.Cui Lexmark.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

October 1999 Statistical Methods for Computer Science Marie desJardins CMSC 601 April 9, 2012 Material adapted.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Sampling: Final and Initial Sample Size Determination
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Ch11 Curve Fitting Dr. Deshi Ye
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
Chapter 8 Estimation: Additional Topics
Psychology 202b Advanced Psychological Statistics, II February 15, 2011.
Classification and risk prediction
Chapter Seventeen HYPOTHESIS TESTING
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Chap 9-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 9 Estimation: Additional Topics Statistics for Business and Economics.
In previous lecture, we dealt with the unboundedness problem of LPM using the logit model. In this lecture, we will consider another alternative, i.e.
Thanks to Nir Friedman, HU
Sampling Distributions & Point Estimation. Questions What is a sampling distribution? What is the standard error? What is the principle of maximum likelihood?
Bootstrap spatobotp ttaoospbr Hesterberger & Moore, chapter 16 1.
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
Choosing Statistical Procedures
Chapter 9 Two-Sample Tests Part II: Introduction to Hypothesis Testing Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social & Behavioral.
5-1 Introduction 5-2 Inference on the Means of Two Populations, Variances Known Assumptions.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
The Triangle of Statistical Inference: Likelihoood
Topics: Statistics & Experimental Design The Human Visual System Color Science Light Sources: Radiometry/Photometry Geometric Optics Tone-transfer Function.
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Chanyoung Park Raphael T. Haftka Paper Helicopter Project.
Chapter 13 – Difference Between Two Parameters Math 22 Introductory Statistics.
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Biostatistics, statistical software VII. Non-parametric tests: Wilcoxon’s signed rank test, Mann-Whitney U-test, Kruskal- Wallis test, Spearman’ rank correlation.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
The Triangle of Statistical Inference: Likelihoood Data Scientific Model Probability Model Inference.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Empirical Likelihood Mai Zhou Dept. of Statistics, University of Kentucky.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Analysis of Chromium Emissions Data Nagaraj Neerchal and Justin Newcomer, UMBC and OIAA/OEI and Mohamed Seregeldin, Office of Air Quality Planning and.
Bayesian Approach For Clinical Trials Mark Chang, Ph.D. Executive Director Biostatistics and Data management AMAG Pharmaceuticals Inc.
T tests comparing two means t tests comparing two means.
Copyright © 2010 Pearson Addison-Wesley. All rights reserved. Chapter 9 One- and Two-Sample Estimation Problems.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Chapter 13 Understanding research results: statistical inference.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Hypothesis Testing. Statistical Inference – dealing with parameter and model uncertainty  Confidence Intervals (credible intervals)  Hypothesis Tests.
Lecture 22 Dustin Lueker.  Similar to testing one proportion  Hypotheses are set up like two sample mean test ◦ H 0 :p 1 -p 2 =0  Same as H 0 : p 1.
DATA ANALYSIS AND MODEL BUILDING LECTURE 7 Prof. Roland Craigwell Department of Economics University of the West Indies Cave Hill Campus and Rebecca Gookool.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
Small area estimation combining information from several sources Jae-Kwang Kim, Iowa State University Seo-Young Kim, Statistical Research Institute July.
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Part Four ANALYSIS AND PRESENTATION OF DATA
Research Methodology Lecture No :25 (Hypothesis Testing – Difference in Groups)
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Linear Mixed Models in JMP Pro
Mai Zhou Dept. of Statistics, University of Kentucky Chengwu Cui
Chapter 9 One- and Two-Sample Estimation Problems.
Chapter 8 Estimation: Additional Topics
Comparing Populations
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
I. Statistical Tests: Why do we use them? What do they involve?
Chapter 24 Comparing Two Means.
S.M.JOSHI COLLEGE, HADAPSAR
Classical regression review
Presentation transcript:

Mathematical Model for the Law of Comparative Judgment in Print Sample Evaluation Mai Zhou Dept. of Statistics, University of Kentucky Luke C.Cui Lexmark International Inc.

The Problem: When evaluating several print samples, pair-wise comparison experiments are often used. Two print samples at a time are judged by a human subject to determine which print sample is “better”. This is repeated with different pairs and different subjects. The resulting data will look like: / / / / / / / / / / 4

How to Summarize the data; Order the print samples in terms of “strength”; Margin of error in the analysis/conclusion. Predict the outcome of future comparisons.

Outline of talk Introduction to Thurstone/Mosteller Model New model, theoretical formulation Var-Cor modeling, Maximum Likelihood Estimation, Likelihood ratio confidence interval New model, application to experimental data Comparisons with classical model, how good is the fit? Discussion

For pairwise comparisons of stimuli i and k, the observable outcomes are the signs of and the outcomes from different pairs are independent. (but within the pair, they may or may not be independent). Assume Where N(, ) denotes the normal distribution.

If we observed the outcomes of many pairs, the log likelihood function is where And is the cdf of the standard normal distribution (available in many software packages).

Where W (or L) is the times stimulus i is deemed better (or worse) than stimulus k in the pair- wise comparisons. The classical model assumes The new model we propose assumes for the variances

Because the human perceptual process is highly adaptive and is at its best when used as a null tester, ie, more sensitive for closely matched stimuli. Thus the variances should be related to how closely the strengths are matched. e.g.

Computation Use software Splus (commercial) or R (Gnu) or Mathcad (commercial) or Matlab (commercial) or SAS (commerical) 1. Define the log likelihood function llk() as a function of the parameters.

2. Maximize the llk() or minimize the negative of llk() by using the optimization functions supplied. In R the optimize functions are: nlm( ) optim( ) In SAS iml we could use function nlptr( )

The parameter values that achieve the maximization (max1) are the estimate of the parameters. Confidence interval of the parameter can be obtained by temporarily fix the value of the parameter at and maximize over the remaining parameters. Suppose it achieved maximum value max2. those values for which max1 – max2 < 3.84/2 is the 95% confidence interval for the parameter.

Example: Colorfulness data Nine print samples were compared. Pairwise experiment, 50 subjects

Models fitted are: 1. Classic model with equal variances. 2. New model

Models fitted are: 2. New model

Differences: (predicted – observed) Model 1

Differences: (predicted – observed) Model 2 with one more para.

Differences: (predicted – observed) Model 1 vs 2

We also fit Bradley-Terry model to the data (use SAS) and the fit is similar to the classic model.

References 1.Peter, G. Engeldrum, Psychometric scaling, A toolkit for imaging system development, Imcotek press. (2000) 2.Torgerson, W.S. Theory and methods of scaling, John Wiley & Sons, Inc. (1958) 3.Bradley, R.A. and Terry, M. E. "Rank analysis of incomplete block design. I. The method of paired comparisons." Biometrika 39, (1952) 4.P. Hall and B. La Scala, Methodology and algorithms of empirical likelihood, International Statistical Review, 58, (1990) R: Updated manuscript: Acknowledgements: We would like to thank Dr. Shaun Love at Lexmark International Inc. for helpful discussions.