Adaptive Methods Research Methods Fall 2010 Tamás Bőhm.

Slides:



Advertisements
Similar presentations
Adaptive Methods Research Methods Fall 2008 Tamás Bőhm.
Advertisements

Dummy Dependent variable Models
Chapter 7 Hypothesis Testing
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
Arc-length computation and arc-length parameterization
Estimation of Means and Proportions
Statistical Analysis SC504/HS927 Spring Term 2008
Commonly Used Distributions
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Giansalvo EXIN Cirrincione unit #3 PROBABILITY DENSITY ESTIMATION labelled unlabelled A specific functional form for the density model is assumed. This.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Sampling: Final and Initial Sample Size Determination
Confidence Intervals This chapter presents the beginning of inferential statistics. We introduce methods for estimating values of these important population.
Thresholds, Weber’s law, Fechner’s three methods Research Methods Fall 2010 Tamás Bőhm.
Maximum likelihood (ML) and likelihood ratio (LR) test
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
GRA 6020 Multivariate Statistics; The Linear Probability model and The Logit Model (Probit) Ulf H. Olsson Professor of Statistics.
Maximum likelihood (ML)
Maximum likelihood (ML) and likelihood ratio (LR) test
PPA 415 – Research Methods in Public Administration Lecture 5 – Normal Curve, Sampling, and Estimation.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
The Binary Logit Model Definition Characteristics Estimation 0.
Maximum likelihood (ML)
Generalized Linear Models
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Single and Multiple Spell Discrete Time Hazards Models with Parametric and Non-Parametric Corrections for Unobserved Heterogeneity David K. Guilkey.
MODELS OF QUALITATIVE CHOICE by Bambang Juanda.  Models in which the dependent variable involves two ore more qualitative choices.  Valuable for the.
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
Overview of Meta-Analytic Data Analysis
Overview Definition Hypothesis
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
1 G Lect 11W Logistic Regression Review Maximum Likelihood Estimates Probit Regression and Example Model Fit G Multiple Regression Week 11.
Model Building III – Remedial Measures KNNL – Chapter 11.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
Model Inference and Averaging
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
PARAMETRIC STATISTICAL INFERENCE
1 Lesson 8: Basic Monte Carlo integration We begin the 2 nd phase of our course: Study of general mathematics of MC We begin the 2 nd phase of our course:
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Chapter 13: Limited Dependent Vars. Zongyi ZHANG College of Economics and Business Administration.
BCS547 Neural Decoding.
Generalized Linear Models (GLMs) and Their Applications.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Logistic regression (when you have a binary response variable)
Maximum likelihood estimators Example: Random data X i drawn from a Poisson distribution with unknown  We want to determine  For any assumed value of.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
The Probit Model Alexander Spermann University of Freiburg SS 2008.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Statistical Modelling
10701 / Machine Learning.
Generalized Linear Models
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
Interval Estimation and Hypothesis Testing
Sampling Distribution of the Mean
Signal detection theory
Diagnostics and Remedial Measures
EE Audio Signals and Systems
Diagnostics and Remedial Measures
Presentation transcript:

Adaptive Methods Research Methods Fall 2010 Tamás Bőhm

Adaptive methods Classical (Fechnerian) methods: stimulus is often far from the threshold  inefficient Adaptive methods: accelerated testing –Modifications of the method of constant stimuli and method of limits

Adaptive methods Classical methods: stimulus values to be presented are fixed before the experiment Adaptive methods: stimulus values to be presented depend critically on preceding responses

Adaptive methods Constituents –Stepping rule: which stimulus level to use next? –Stopping criterion: when to finish the session? –What is the final threshold estimate? Performance –Bias: systematic error –Precision: related to random error –Efficiency: # of trials needed for a specific precision; measured by the sweat factor

Notations X n stimulus level at trial n Z n response at trial n Z n = 1detected / correct Z n = 0not detected / incorrect φtarget probability absolute threshold: φ = 50% difference threshold: φ = 75% 2AFC: φ = 50% + 50% / 2 = 75% 4AFC: φ = 25% + 75% / 2 = 62.5% x φ threshold

Adaptive methods Classical methods: stimulus values to be presented are fixed before the experiment Adaptive methods: stimulus values to be presented depend critically on preceding responses X n+1 = f(φ, n, Z n, X n, Z n-1, X n-1,…, Z 1, X 1 )

Adaptive methods Nonparametric methods: –No assumptions about the shape of the psychometric function –Can measure threshold only Parametric methods: –General form of the psychometric function is known, only its parameters (threshold and slope) need to be measured –If slope is also known: measure only threshold

Nonparametric adaptive methods Staircase method (aka. truncated method of limits, simple up-down) Transformed up-down method Nonparametric up-down method Weighted up-down method Modified binary search Stochastic approximation Accelerated stochastic approximation PEST and More Virulent PEST

Staircase method Stepping rule: X n+1 = X n - δ(2Z n - 1) –fixed step size δ –if response changes: direction of steps is reversed Stopping criterion: after a predetermined number of reversals Threshold estimate: average of reversal points (mid-run estimate) Converges to φ = 50%  cannot be used for e.g. 2AFC

Transformed up-down method Improvement of the simple up-down (staircase) method X n+1 depends on 2 or more preceding responses –E.g.1-up/2-down or 2- step rule: Increase stimulus level after each incorrect response Decrease only after 2 correct responses φ = 70.7% Threshold: mid-run estimate 8 rules for 8 different φ values (15.9%, 29.3%, 50%, 70.7%, 79.4%, 84.1%) reversal points

Nonparametric up-down method Stepping rule:X n+1 = X n - δ(2Z n S φ - 1) –S φ : random number p(S φ =1) = 1 / 2φ p(S φ =0) = 1 – (1 / 2φ) –After a correct answer: stimulus decreased with p = 1 / 2φ stimulus increased with p = 1 - (1 / 2φ) –After an incorrect answer: stimulus increased Can converge to any φ ≥ 50%

Nonparametric up-down method

Weighted up-down method Different step sizes for upward (δ up ) and downward steps (δ down )

Modified binary search ‘Divide and conquer’ Stimulus interval containing the threshold is halved in every step (one endpoint is replaced by the midpoint) Stopping criterion: a lower limit on the step size Threshold estimate: last tested level Heuristic, no theoretical foundation Figure from Sedgewick & Wayne

Stochastic approximation A theoretically sound variant of the modified binary search Stepping rule: –c: initial step size –Stimulus value increases for correct responses, decreases for incorrect ones –If φ = 50%: upward and downward steps are equal; otherwise asymmetric –Step size (both upward and downward) decreases from trial to trial Can converge to any φ

Stochastic approximation

Accelerated stochastic approximation Stepping rule: –First 2 trials: stochastic approximation –n > 2: step size is changed only when response changes (m reversals : number of reversals) Otherwise the same as stochastic approximation Less trials than stochastic approximation

Accelerated stochastic approximation

Parameter Estimation by Sequential Testing (PEST) Sequential testing: –Run multiple trials at the same stimulus level x –If x is near the threshold, the expected number of correct responses m c after n x presentations will be around φn x  the stimulus level is changed if m c is not in φn x ± w –w: deviation limit; w=1 for φ=75% If the stimulus level needs to be changed: step size determined by a set of heuristic rules Variants: MOUSE, RAT, More Virulent PEST

Adaptive methods Nonparametric methods: –No assumptions about the shape of the psychometric function –Can measure threshold only Parametric methods: –General form of the psychometric function is known, only its parameters (threshold and slope) need to be measured –If slope is also known: measure only threshold

Parametric adaptive methods A template for the psychometric function is chosen: –Cumulative normal –Logistic –Weibull –Gumbel

Parametric adaptive methods Only the parameters of the template need to be measured: –Threshold –Slope

Fitting the psychometric function 1.Linearization (inverse transformation) of data points Inverse cumulative normal (probit) Inverse logistic (logit)

Fitting the psychometric function 2.Linear regression 3.Transformation of regression line parameters X-intercept & linear slopeThreshold & logistic slope

Contour integration experiment D = 2 slope = -0.6 D = 65 slope = 0.3

Contour integration experiment 5-day perceptual learning

Adaptive probit estimation Short blocks of method of constant stimuli Between blocks: threshold and slope is estimated (psychometric function is fitted to the data) and stimulus levels adjusted accordingly –Assumes a cumulative normal function  probit analysis Stopping criterion: after a fixed number of blocks Final estimate of threshold and slope: re-analysis of all the responses

Adaptive probit estimation Start with an educated guess of the threshold and slope In each block: 4 stimulus levels presented 10 times each After each block: threshold ( ) and slope ( ) is estimated by probit analysis of the responses in block Stimulus levels for the next block are adjusted accordingly –Estimated threshold and slope applied only through correction factors  inertia

Measuring the threshold only Function shape (form & slope) is predetermined by the experimenter Only the position along the x-axis (threshold) needs to be measured Iteratively estimating the threshold and adapting the stimulus levels Two ways to estimate the threshold: –Maximum likelihood (ML) –Bayes’ estimation QUEST, BEST PEST, ML-TEST, Quadrature Method, IDEAL, YAAP, ZEST

Maximum likelihood estimation Construct the psychometric function with each possible threshold value Calculate the probability of the responses with each threshold value (likelihood) Choose the threshold value for which the likelihood is maximal (i.e. the psychometric function that is the most likely to produce such responses)

Bayes’ estimation Prior information is also used –Distribution of the threshold in the population (e.g. from a survey of the literature) –The experimenter’s beliefs about the threshold a priori distribution of the threshold values of the psychometric functions at the tested stimulus levels