Image Analysis, Random Fields and Dynamic MCMC By Marc Sobel.

Slides:



Advertisements
Similar presentations
Bayesian inference Jean Daunizeau Wellcome Trust Centre for Neuroimaging 16 / 05 / 2008.
Advertisements

Bayesian Belief Propagation
Bayesian Estimation in MARK
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Smoothing 3D Meshes using Markov Random Fields
Markov random field Institute of Electronics, NCTU
BAYESIAN INFERENCE Sampling techniques
CS774. Markov Random Field : Theory and Application Lecture 16 Kyomin Jung KAIST Nov
J. Mike McHugh,Janusz Konrad, Venkatesh Saligrama and Pierre-Marc Jodoin Signal Processing Letters, IEEE Professor: Jar-Ferr Yang Presenter: Ming-Hua Tang.
1 On the Statistical Analysis of Dirty Pictures Julian Besag.
The Markov property Discrete time: A time symmetric version: A more general version: Let A be a set of indices >k, B a set of indices
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
1 Markov random field: A brief introduction Tzu-Cheng Jen Institute of Electronics, NCTU
Thanks to Nir Friedman, HU
Algorithm Evaluation and Error Analysis class 7 Multiple View Geometry Comp Marc Pollefeys.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Lecture II-2: Probability Review
Robin McDougall, Ed Waller and Scott Nokleby Faculties of Engineering & Applied Science and Energy Systems & Nuclear Science 1.
Crash Course on Machine Learning
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
MCMC: Particle Theory By Marc Sobel. Particle Theory: Can we understand it?
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 7th~10th Belief propagation Appendix Kazuyuki Tanaka Graduate School of Information.
Techniques for Estimating Layers from Polar Radar Imagery Jerome E. Mitchell, Geoffrey C. Fox, and David J. Crandall :: CReSIS NSF Site Visit :: School.
EM and expected complete log-likelihood Mixture of Experts
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
A Markov Random Field Model for Term Dependencies Donald Metzler W. Bruce Croft Present by Chia-Hao Lee.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Markov Random Fields Probabilistic Models for Images
First topic: clustering and pattern recognition Marc Sobel.
The Dirichlet Labeling Process for Functional Data Analysis XuanLong Nguyen & Alan E. Gelfand Duke University Machine Learning Group Presented by Lu Ren.
Xu Huaping, Wang Wei, Liu Xianghua Beihang University, China.
An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.
14 October, 2010LRI Seminar 2010 (Univ. Paris-Sud)1 Statistical performance analysis by loopy belief propagation in probabilistic image processing Kazuyuki.
1 Markov Random Fields with Efficient Approximations Yuri Boykov, Olga Veksler, Ramin Zabih Computer Science Department CORNELL UNIVERSITY.
National Alliance for Medical Image Computing Segmentation Foundations Easy Segmentation –Tissue/Air (except bone in MR) –Bone in CT.
1 Markov random field: A brief introduction (2) Tzu-Cheng Jen Institute of Electronics, NCTU
An Introduction to Kalman Filtering by Arthur Pece
MCMC (Part II) By Marc Sobel. Monte Carlo Exploration  Suppose we want to optimize a complicated distribution f(*). We assume ‘f’ is known up to a multiplicative.
CS 8751 ML & KDDData Clustering1 Clustering Unsupervised learning Generating “classes” Distance/similarity measures Agglomerative methods Divisive methods.
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Lecture 2: Statistical learning primer for biologists
Flat clustering approaches
Spectrum Sensing In Cognitive Radio Networks
Graduate School of Information Sciences, Tohoku University
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
6.4 Random Fields on Graphs 6.5 Random Fields Models In “Adaptive Cooperative Systems” Summarized by Ho-Sik Seok.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
STA347 - week 91 Random Vectors and Matrices A random vector is a vector whose elements are random variables. The collective behavior of a p x 1 random.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Regularization of energy-based representations Minimize total energy E p (u) + (1- )E d (u,d) E p (u) : Stabilizing function - a smoothness constraint.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Ch 6. Markov Random Fields 6.1 ~ 6.3 Adaptive Cooperative Systems, Martin Beckerman, Summarized by H.-W. Lim Biointelligence Laboratory, Seoul National.
Markov random fields. The Markov property Discrete time: A time symmetric version: A more general version: Let A be a set of indices >k, B a set of indices.
6.8 Maximizer of the Posterior Marginals 6.9 Iterated Conditional Modes of the Posterior Distribution Jang, HaYoung.
Biointelligence Laboratory, Seoul National University
MCMC Output & Metropolis-Hastings Algorithm Part I
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
Classification of unlabeled data:
Markov Random Fields with Efficient Approximations
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Binarization of Low Quality Text Using a Markov Random Field Model
Statistics and Shape Analysis
Markov Random Fields for Edge Classification
Outline Texture modeling - continued Markov Random Field models
Presentation transcript:

Image Analysis, Random Fields and Dynamic MCMC By Marc Sobel

A field!!!!!!!!

Random Fields  Random Fields RF consist in collections of points P={p} and neighborhoods {N p } of points. (Neighborhoods N p do not contain p) The field imposes ‘label’ values f={f[p]} points. We use the notation f[S] for the label values imposed on a set S Random Fields have one central property which is closely related to the markov property:

Reasoning: Hammersely-Clifford Theorem  Under certain assumptions, assuming the points can be enumerated by p 1,…,p N we have that: (the distribution can be generated from these conditionals)

Gibbs Random Field  Gibbs random fields are characterized by  Where cε  are cliques. Cliques are contained in neighborhoods:  ⊂   For example, if cliques c are all pairs, we could put : 

Gibbs=Markov!!!!!!  Under Gibbs, conditioning is on neighborhoods:  But, the term,  Cancels in numerator  and denominator  giving the result

Examples of Random Fields  Automodels: all cliques have one or two members.  Autobinomial models: How to biuld a k color map: labels are 0,1,…,k. Neighborhoods are of size M.  Autologistic model (i.e., Model which imposes energy 1 when contiguous elements are different and -1 otherwise). 

A Metropolis Hastings update for autologistic field models  1) Propose a flip at a randomly selected point p.  2) The move probability is:

The 1-d autologistic  The 1-d autologistic is:  The effect of the prior is to smooth out the results.

The 2-d autologistic or Ising Model  In a 2-d setting we update using:

Example: The Ising Model: Each rectangle below is a field configuration f: black=1 and white=-1. Color results from multiple label values

Extensions: Segmentation  Start with a map Y (over a 2d grid).  Assume we would like to distinguish which points in the map are importand and which are background.  Devise an ising field model prior which captures the importand points of the maps and downweights the others. E.g.,

Extensions (concluded)  So, minimizing the potential contains a ‘magnetic field’ (based on the first term) and an ‘external field’ based on the second term.  Other extensions are to Line processes, image reconstruction, texture representation.

Random Field Priors: The Ising model or autologistic model: (Metropolis Hastings updates): Temperature=5; at time t= Note the presence of more ‘islands’.

Random Field Priors: The Ising model. (Metropolis Hastings updates):Temperature=.005; Note the presence of fewer islands.

Generalized Ising models: Mean Field Equation  The energy is:  What is the ‘impact of this prior’? Use mean field equations to get the closest possible prior (in KLD) which makes the field points mutually independent.

Generalized Field: Note the Swiss Cheese aspect. Temperature=10.

Mean Field Equation  The mean field equation minimizes:  For distributions Q which make points mutually independent. For the generalized field model, the mean field equation is:

Mean Field Approximation to the General Ising Field at temperature T=10. We simulate from the mean field prior. We retain the swiss cheese but lose the islands.

Gaussian Process  Autonormal models: If labels are real numbers, (i.e., we are trying to biuld a picture with many different grey levels):

Gaussian Processes For Gaussian processes, the covariance, Cov(f[p],f[p’])= Σ β p’,p’’ cov(f[p’],f[p’’]) + σ 2 ; This gives the Yule-Walker equation: COV=B*COV+I; or COV -1 =(I-B)/σ 2 ; So the likelihood is given by,

Gaussian Processes  The likelihood is gaussian with mean μ and inverse covariance matrix I-B;  Example: assume a likelihood, centered at i+j. Assume a gaussian process prior.

Posterior Distribution for the Gaussian Model

Gaussian field at time t=20,000 with conditional prior variance=.01. Mesh is over a realization of μ. Note how smooth the mesh is:

Maximum Aposteriori Estimates

MAP Estimator with prior variance =.5

Maximum Aposteriori Estimate with prior variance =.01

Smoothness Priors  Suppose we observe data with prior,

Smoothness priors  The smoothness prior π 1 has the effect of imposing a small ‘derivative’ on the field.  The smoothness prior π 2 has the effect of imposing a small curvature on the field.

Smoothness Priors  Smoothness priors have the same kind of impact as choosing a function which minimizes the ‘loss’,  Assume the likelihood

Data = -5 below 50 and Data=5 above 50. Conditional prior variance is.5

Data = -5 below 50 and Data=5 above 50. Conditional prior variance is.005