Bayesian evidence for visualizing model selection uncertainty Gordon L. Kindlmann

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Pattern Recognition and Machine Learning
Pattern Recognition and Machine Learning
Tutorial on Bayesian Techniques for Inference A.Asensio Ramos Instituto de Astrofísica de Canarias.
Learning from spectropolarimetric observations A. Asensio Ramos Instituto de Astrofísica de Canarias aasensio.github.io/blog.
Bayesian Estimation in MARK
MEG/EEG Inverse problem and solutions In a Bayesian Framework EEG/MEG SPM course, Bruxelles, 2011 Jérémie Mattout Lyon Neuroscience Research Centre ? ?
1 Methods of Experimental Particle Physics Alexei Safonov Lecture #21.
Integration of sensory modalities
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Chapter 4: Linear Models for Classification
Diffusion Tensor MRI And Fiber Tacking Presented By: Eng. Inas Yassine.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Bayesian Learning Rong Jin. Outline MAP learning vs. ML learning Minimum description length principle Bayes optimal classifier Bagging.
Probabilistic inference
Bayes Rule How is this rule derived? Using Bayes rule for probabilistic inference: –P(Cause | Evidence): diagnostic probability –P(Evidence | Cause): causal.
Constraining Astronomical Populations with Truncated Data Sets Brandon C. Kelly (CfA, Hubble Fellow, 6/11/2015Brandon C. Kelly,
Decision making as a model 7. Visual perception as Bayesian decision making.
Neural Networks Marco Loog.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Machine Learning CMPT 726 Simon Fraser University
Bayesian Learning Rong Jin.
Computer vision: models, learning and inference
Rician Noise Removal in Diffusion Tensor MRI
Hydrologic Statistics
Computer vision: models, learning and inference Chapter 6 Learning and Inference in Vision.
Bayesian Model Comparison and Occam’s Razor Lecture 2.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Informatics and Mathematical Modelling / Intelligent Signal Processing 1 EUSIPCO’09 27 August 2009 Tuning Pruning in Sparse Non-negative Matrix Factorization.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
EM and expected complete log-likelihood Mixture of Experts
DTU Medical Visionday May 27, 2009 Generative models for automated brain MRI segmentation Koen Van Leemput Athinoula A. Martinos Center for Biomedical.
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
28 February, 2003University of Glasgow1 Cluster Variation Method and Probabilistic Image Processing -- Loopy Belief Propagation -- Kazuyuki Tanaka Graduate.
Sample variance and sample error We learned recently how to determine the sample variance using the sample mean. How do we translate this to an unbiased.
Part II: Model Class Selection Given: Dynamic data from system and set of candidate model classes where each model class defines a set of possible predictive.
References: [1]S.M. Smith et al. (2004) Advances in functional and structural MR image analysis and implementation in FSL. Neuroimage 23: [2]S.M.
14 October, 2010LRI Seminar 2010 (Univ. Paris-Sud)1 Statistical performance analysis by loopy belief propagation in probabilistic image processing Kazuyuki.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
INTRODUCTION TO Machine Learning 3rd Edition
July 11, 2006Bayesian Inference and Maximum Entropy Probing the covariance matrix Kenneth M. Hanson T-16, Nuclear Physics; Theoretical Division Los.
Sequential Monte-Carlo Method -Introduction, implementation and application Fan, Xin
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston Ch. 5 Bayesian Treatment of Neuroimaging Data Will Penny and Karl Friston 18.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Regression Analysis1. 2 INTRODUCTION TO EMPIRICAL MODELS LEAST SQUARES ESTIMATION OF THE PARAMETERS PROPERTIES OF THE LEAST SQUARES ESTIMATORS AND ESTIMATION.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Machine Learning Expectation Maximization and Gaussian Mixtures CSE 473 Chapter 20.3.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Machine Learning Expectation Maximization and Gaussian Mixtures CSE 473 Chapter 20.3.
Statistical Concepts Basic Principles An Overview of Today’s Class What: Inductive inference on characterizing a population Why : How will doing this allow.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Bayesian Framework Finding the best model Minimizing model complexity
Statistical-Mechanical Approach to Probabilistic Image Processing -- Loopy Belief Propagation and Advanced Mean-Field Method -- Kazuyuki Tanaka and Noriko.
Graduate School of Information Sciences, Tohoku University
Special Topics In Scientific Computing
Graduate School of Information Sciences, Tohoku University
Estimating the Spatial Sensitivity Function of A Light Sensor N. K
CSCI 5822 Probabilistic Models of Human and Machine Learning
Graduate School of Information Sciences, Tohoku University
M/EEG Statistical Analysis & Source Localization
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Will Penny Wellcome Trust Centre for Neuroimaging,
Presentation transcript:

Bayesian evidence for visualizing model selection uncertainty Gordon L. Kindlmann

Basic message There many different kinds of uncertainty (Rheingans “Ways of not knowing”, Monday) There’s at least one more kind of uncertainty Based on an empirical view of data & visualization (not sure if relevant for simulation studies?) Goal here is to: Define this kind of uncertainty Argue for its relevance Get feedback from you Work initiated by Maxwell Shron (now data scientist for OkCupid); collaboration with Thomas Schultz

frequency One flavor of uncertainty Variance in measurement of scalar (e.g. length) uncertainty in length due to the measurement process itself Potter et al. "Visualizing Summary Statistics and Uncertainty". Eurovis 2010 pp. 823–832 Pöthkow et al. "Probabilistic Marching Cubes". Eurovis 2011 pp length 20”22”

Second flavor of uncertainty Variance in parameters of model of data Model: mathematical representation of a hypothesis, parameterized by physically meaningful degrees-of-freedom, to predict measurements Doppler shift in star light color from planet orbit Decrease in star light intensity due to planet transit Finding planets around other stars Jan 2011: Kepler telescope detects rocky planet “Kepler-10b”

Second flavor of uncertainty Variance in parameters of model of data => Parameters of models of empirical data are fundamental quantities in the scientific method DK Jones, “Determining and visualizing uncertainty in estimates of fiber orientation from diffusion tensor MRI” Magnetic Resonance in Medicine 49:7-12 (2003) Rheingans & desJardins, “Visualizing High- Dimensional Predictive Model Quality.” Visualization 2000, pp,

L (pos,height) position height Know something about noise => compute likelihood L of the data given particular model parameters Second flavor of uncertainty Variance in parameters of model of data Synthetic data and low-DOF model DataModel (2 DOF) peak position peak height = 1/width Maximum Likelihood Estimate

Proposed new flavor of uncertainty 0 Data L (pos,height) 2 Data L (pos,height) 4 Data L (pos,height) 6 Data L (pos,height) 8 Data L (pos,height) A Data L (pos,height) position height 1 peak: 2 DOF 2 peaks: 4 DOF ???? Ambiguity in choice of model to describe data

Goals of visualizing model ambiguity 1) Understand quality/sufficiency of set of models in novel and complex imaging modalities Modern imaging produces multiple values per-voxel Discover spatial/anatomic structure of where models are descriptive, and where not Data Model A Model B Model C Model D

Algorithm Answer D Answer C Answer B Answer A Goals of visualizing model ambiguity Data Model A Model B Model C Model D 1) Understand quality/sufficiency of set of models in novel and complex imaging modalities Modern imaging produces multiple values per-voxel Discover spatial/anatomic structure of where models are descriptive, and where not 2) Characterize stability of visualization/analysis with respect to changes in model

D Model Fitting Diffusion Tensors from DWI data Single Tensor Model (Basser et al. 1994) : S0S0 gigi SiSi Diffusion-weighted imaging (DWI) experiment Scalar invariants of tensor: Bulk mean diffusivity “MD” Fractional Anisotropy “FA”

How to quantify model quality? Can use error (residual) in fit More DOF => better fit, but less explanatory One method: Bayesian Model Inference Naturally implements Occam’s Razor Cleanly includes probabilistic noise model Various schemes for penalizing high DOF

Bayes Visualized U niverse A B AB ⇒

Posterior Likelihood Prior D position height Bayes for fitting (one, fixed) model Find parameters w of model M i that maximize posterior probability of the measured data D Friman et al. “A Bayesian approach for stochastic white matter tractography” IEEE TMI 25: (2006) Maximum A Posteriori Estimate

PriorEvidence Bayes for evaluating a model Want to quantify: how plausible is model M i, as a whole, given the data? Evidence Evidence: integral of data likelihood over model’s entire parameter space (not specific to single fit) Likelihood Uncertainty in model choice (of M 1 vs M 2 ) is quantified by comparing evidence(M 1 ), evidence(M 2 ) Perhaps Sum(evidence) shows gaps in models? constant

Evidence plots over 1-D domain f(y) = 0 Evidence(x) data(x) = f(y) x 0

Evidence plots over 1-D domain f(y) = 0 f(y) = C Evidence(x) data(x) = f(y) x 0 1

Evidence plots over 1-D domain f(y) = H*bump(y) f(y) = C f(y) = H*bump(y-P) Evidence(x) data(x) = f(y) x 1 1 2

Evidence plots over 1-D domain f(y) = H*bump(y) f(y) = C f(y) = H*bump(y-P) Evidence(x) data(x) = f(y) x 1 2 1

Discussion New opportunity for visualization to help understand statistics & add scientific value Data & models complex; can see how data can be explained, and where it can’t Can show exactly when simple models suffice, backed up with statistical theory Open research question: how to visualize the multiple scalar fields generated by computing per-pixel model evidences? Thank you!