Are you being blinded by statistics? The structure of the retinal nerve fibre layer.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Sampling Distributions
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Continuous Probability Distributions.  Experiments can lead to continuous responses i.e. values that do not have to be whole numbers. For example: height.
Measures of Dispersion
Calculating & Reporting Healthcare Statistics
Point estimation, interval estimation
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Chapter Eighteen MEASURES OF ASSOCIATION
Lecture 20 Simple linear regression (18.6, 18.9)
Introduction to Educational Statistics
Archimedean Copulas Theodore Charitos MSc. Student CROSS.
The Lognormal Distribution
Lecture II-2: Probability Review
The arithmetic mean of a variable is computed by determining the sum of all the values of the variable in the data set divided by the number of observations.
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
Inferential Statistics: SPSS
3 CHAPTER Cost Behavior 3-1.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 12 Describing Data.
Chap 6-1 Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall Chapter 6 The Normal Distribution Business Statistics: A First Course 6 th.
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
Overview Summarizing Data – Central Tendency - revisited Summarizing Data – Central Tendency - revisited –Mean, Median, Mode Deviation scores Deviation.
PROBABILITY & STATISTICAL INFERENCE LECTURE 3 MSc in Computing (Data Analytics)
Traffic Modeling.
© 2008 Brooks/Cole, a division of Thomson Learning, Inc. 1 Chapter 4 Numerical Methods for Describing Data.
On project probabilistic cost analysis from LHC tender data Ph. Lebrun CERN, Geneva, Switzerland TILC’09, Tsukuba, Japan April 2009.
Instrumentation (cont.) February 28 Note: Measurement Plan Due Next Week.
Descriptive Statistics: Numerical Methods
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Measures of Dispersion & The Standard Normal Distribution 9/12/06.
1 Section 9-4 Two Means: Matched Pairs In this section we deal with dependent samples. In other words, there is some relationship between the two samples.
Regression Models Residuals and Diagnosing the Quality of a Model.
Copyright © 2011 Pearson Education, Inc. The Simple Regression Model Chapter 21.
Regression Chapter 16. Regression >Builds on Correlation >The difference is a question of prediction versus relation Regression predicts, correlation.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
1 Statistical Analysis – Descriptive Statistics Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND.
© Copyright McGraw-Hill 2000
Numerical Measures of Variability
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 21 The Simple Regression Model.
Slide 1 Copyright © 2004 Pearson Education, Inc..
1 Chapter 4 Numerical Methods for Describing Data.
Data Analysis.
Sampling and estimation Petter Mostad
Psychology 202a Advanced Psychological Statistics September 10, 2015.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Probability distributions
Variability Introduction to Statistics Chapter 4 Jan 22, 2009 Class #4.
Session V: The Normal Distribution Continuous Distributions (Zar, Chapter 6)
Chapter 6: Descriptive Statistics. Learning Objectives Describe statistical measures used in descriptive statistics Compute measures of central tendency.
Copyright © 2016 Brooks/Cole Cengage Learning Intro to Statistics Part II Descriptive Statistics Intro to Statistics Part II Descriptive Statistics Ernesto.
CHAPTER 3 – Numerical Techniques for Describing Data 3.1 Measures of Central Tendency 3.2 Measures of Variability.
1 Estimation Chapter Introduction Statistical inference is the process by which we acquire information about populations from samples. There are.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Central Bank of Egypt Basic statistics. Central Bank of Egypt 2 Index I.Measures of Central Tendency II.Measures of variability of distribution III.Covariance.
CHAPTER 15: THE NUTS AND BOLTS OF USING STATISTICS.
Best in Market Pricing. What is Best in Market Pricing ? An extension of parametric modeling for negotiating lowest pricing for a statement of work consisting.
Parameter, Statistic and Random Samples
BAE 5333 Applied Water Resources Statistics
STATISTICS Random Variables and Distribution Functions
Multiple Regression.
Experimental Power Graphing Program
FOUR MOMENTS: Ways to Describe a (Univariate I/R) Distribution
Statistical Methods For Engineers
Evaluation of measuring tools: reliability
Integration of sensory modalities
FOUR MOMENTS: Ways to Describe a (Univariate I/R) Distribution
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

Are you being blinded by statistics?

The structure of the retinal nerve fibre layer

A greyscale showing the visual field for a glaucomatous eye Blind Spot Defect

The principle of filtering – reduce the noise, but preserve the information of interest

Effect of the noise-reduction filter on the greyscale

Noise Currently measured by test-retest variability Currently assumed to be normally distributed

The Principle: Noise = Raw - Filtered

Why this method? Current estimates of the noise are based on test-retest variability. This greatly reduces the amount of data available It imposes more cost and inconvenience on the clinician and patient It is actually estimating double the noise!

Analysing the noise Calculate the noise at each point of each visual field Separate out the noise estimates according to the filtered sensitivity (in 0.5dB bands) Estimate the empirical distribution characteristics at each sensitivity

Empirical pdf of the raw sensitivity when the filtered sensitivity is between 28.25dB and 28.75dB

Empirical pdf of the raw sensitivity when the filtered sensitivity is between 9.25dB and 9.75dB

How does the noise change? As the sensitivity decreases; The mean noise changes The variance increases exponentially The negative skewness increases The censoring at zero becomes apparent

Variance Because of the censoring at 0dB, the variance appears lower than it should be at low sensitivities So base model on noise estimates whose filtered sensitivity is greater than 16dB Model: (R 2 =99.3%)

Skewness Again, because of the censoring at 0dB, the negative skewness disappears (and becomes positive) at low sensitivities So base model on noise estimates whose filtered sensitivity is greater than 16dB Model: (R 2 =87.8%)

Kurtosis Kurtosis is even more affected by the censoring There is no discernable pattern for filtered sensitivities greater than 16dB So use average value instead:

Mean The mean should be less affected by the censoring, so values down to around 10dB should be reliable. Why does it change at all? Because the filter will naturally make sensitivities less extreme. Based on the sensitivities above the noticeable change in pattern (i.e. S>8.5dB), a damped oscillation model fits well.

(crosses represent the empirical mean at each sensitivity; dots represent the modelled mean at the same point)

So, we have values for: Mean Variance Skewness Kurtosis of the noise at any given sensitivity.

Pearson Distributions The Pearson family consists of solutions to which produce well-defined density functions. Given the first four moments of an empirical distribution, you can produce a Pearson density function consistent with those moments.

Pearson Type IV Distributions There are seven types of distribution, depending on the values of the third and fourth moments. For our data, provided the sensitivity is at least 8.5dB, we have a Type IV distribution; has no real roots. This gives a density function of the form:

Steps to the distribution: 1.Predict the mean, variance, skewness and kurtosis of the distribution 2.Use these to calculate the Pearson constants 3.Use numerical integration to obtain a well-defined density function 4.Adjust the mean of the distribution so that it equals the predicted mean

So, we can model the noise distribution - or equivalently the distribution of the raw sensitivity - at any given input sensitivity. This model will be compared with the empirical distribution, and with a normal model.

Filtered Sensitivity between 30.25dB and 30.75dB

Filtered Sensitivity between 10.25dB and 10.75dB

As the sensitivity increases, the pdf of the noise distribution changes. Empirical distribution Modelled distribution Normal distribution

How good is the fit? Calculate Pearson chi-squared statistics: Compare with the statistics when the noise is modelled by a normal distribution, with mean zero and the same variance

Pearson Chi-squared statistics when the data is fit using the model (circles) and using a normal distribution (crosses)

So the fit is much improved. The model even performs better when we are extrapolating our predictions for the variance, skewness and kurtosis downwards below 16dB.

Conclusion: Our new model is better than the current one!

Future Work: Test the model against independent data Extend the model to one valid below 8dB Use the model to simulate visual fields for research in other areas, such as improving testing strategies

The End