(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.

Slides:



Advertisements
Similar presentations
Bayesian inference of normal distribution
Advertisements

Pattern Recognition and Machine Learning
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Rob Fergus Courant Institute of Mathematical Sciences New York University A Variational Approach to Blind Image Deconvolution.
ECE 472/572 - Digital Image Processing Lecture 8 - Image Restoration – Linear, Position-Invariant Degradations 10/10/11.
Visual Recognition Tutorial
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Dimensional reduction, PCA
Independent Component Analysis (ICA) and Factor Analysis (FA)
Expectation Maximization for GMM Comp344 Tutorial Kai Zhang.
Visual Recognition Tutorial
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Computer vision: models, learning and inference
EE462 MLCV 1 Lecture 3-4 Clustering (1hr) Gaussian Mixture and EM (1hr) Tae-Kyun Kim.
A Unifying Review of Linear Gaussian Models
Particle Filtering. Sensors and Uncertainty Real world sensors are noisy and suffer from missing data (e.g., occlusions, GPS blackouts) Use sensor models.
Multidimensional Data Analysis : the Blind Source Separation problem. Outline : Blind Source Separation Linear mixture model Principal Component Analysis.
Review of Probability.
Image Representation Gaussian pyramids Laplacian Pyramids
School of Electrical & Computer Engineering Image Denoising Using Gaussian Scale Mixtures in the Wavelet domain Alex Cunningham Ben Clarke Dy narath Eang.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Latent Variable Models Christopher M. Bishop. 1. Density Modeling A standard approach: parametric models  a number of adaptive parameters  Gaussian.
Machine Learning Queens College Lecture 3: Probability and Statistics.
CAP5415: Computer Vision Lecture 4: Image Pyramids, Image Statistics, Denoising Fall 2006.
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
Review for Exam I ECE460 Spring, 2012.
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
School of Electrical & Computer Engineering Image Denoising Using Steerable Pyramids Alex Cunningham Ben Clarke Dy narath Eang ECE November 2008.
Sample variance and sample error We learned recently how to determine the sample variance using the sample mean. How do we translate this to an unbiased.
Estimating the Likelihood of Statistical Models of Natural Image Patches Daniel Zoran ICNC – The Hebrew University of Jerusalem Advisor: Yair Weiss CifAR.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
Math 3360: Mathematical Imaging Prof. Ronald Lok Ming Lui Department of Mathematics, The Chinese University of Hong Kong Lecture 11: Types of noises.
EE565 Advanced Image Processing Copyright Xin Li Image Denoising: a Statistical Approach Linear estimation theory summary Spatial domain denoising.
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Lecture 2: Statistical learning primer for biologists
Principal Component Analysis (PCA)
Adaptive Wavelet Packet Models for Texture Description and Segmentation. Karen Brady, Ian Jermyn, Josiane Zerubia Projet Ariana - INRIA/I3S/UNSA June 5,
Tracking with dynamics
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
Random Processes Gaussian and Gauss-Markov processes Power spectrum of random processes and white processes.
Intro to Probability Slides from Professor Pan,Yan, SYSU.
Review of statistical modeling and probability theory Alan Moses ML4bio.
The Chinese University of Hong Kong
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Joo-kyung Kim Biointelligence Laboratory,
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Biointelligence Laboratory, Seoul National University
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
Directional Multiscale Modeling of Images
Classification of unlabeled data:
The Chinese University of Hong Kong
Special Topics In Scientific Computing
Image Analysis Image Restoration.
Image Denoising in the Wavelet Domain Using Wiener Filtering
Latent Variables, Mixture Models and EM
Image Segmentation Techniques
Wavelet-Based Denoising Using Hidden Markov Models
Wavelet-Based Denoising Using Hidden Markov Models
A Tutorial on Bayesian Speech Feature Enhancement
EE513 Audio Signals and Systems
Pattern Recognition and Machine Learning
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
Presentation transcript:

(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation (shown zero mean for notational convenience) : If we rotate coordinates to the Fourier basis, the covariance matrix in that basis will be diagonal. So in that model, each Fourier transform coefficient is an independent Gaussian random variable of covariance Image pixels Inverse covariance matrix

Power spectra of typical images Experimentally, the power spectrum as a function of Fourier frequency is observed to follow a power law.

Random draw from Gaussian spectral model

Noise removal (in frequency domain), under Gaussian assumption Variance of white, Gaussian additive noise Observed Fourier component Estimated Fourier component Power law prior probability on estimated Fourier component Setting to zero the derivative of the the log probability of X gives an analytic form for the optimal estimate of X (or just complete the square): Posterior probability for X

Noise removal, under Gaussian assumption original With Gaussian noise of std. dev added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR= (try to ignore JPEG compression artifacts from the PDF file)

(2) The wavelet marginal model Histogram of wavelet coefficients, c, for various images. Parameter determining width of distribution Parameter determining peakiness of distribution Wavelet coefficient value

Random draw from the wavelet marginal model

And again something that is reminiscent of operations found in V1…

An application of image pyramids: noise removal

Image statistics (or, mathematically, how can you tell image from noise?) Noisy image

Clean image

Pixel representation, image histogram

Pixel representation, noisy image histogram

bandpassed representation image histogram

Pixel domain noise image and histogram

Bandpass domain noise image and histogram

Noise-corrupted full-freq and bandpass images But want the bandpass image histogram to look like this

P(x, y) = P(x|y) P(y) so P(x|y) P(y) = P(y|x) P(x) P(x, y) = P(x|y) P(y) so P(x|y) P(y) = P(y|x) P(x) and P(x|y) = P(y|x) P(x) / P(y) Bayes theorem The parameters you want to estimate What you observe Prior probability Likelihood function Constant w.r.t. parameters x. P(x, y) = P(x|y) P(y) By definition of conditional probability Using that twice

P(x) Bayesian MAP estimator for clean bandpass coefficient values Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) P(y|x) P(x|y) P(y|x) y y = 25

Bayesian MAP estimator Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) y P(y|x) P(x|y) y = 50

Bayesian MAP estimator Let x = bandpassed image value before adding noise. Let y = noise-corrupted observation. By Bayes theorem P(x|y) = k P(y|x) P(x) y P(y|x) P(x|y) y = 115

P(x) P(y|x) y y = 25 P(x|y) y P(y|x) P(x|y) y = 115 For small y: probably it is due to noise and y should be set to 0 For large y: probably it is due to an image edge and it should be kept untouched

MAP estimate,, as function of observed coefficient value, y Simoncelli and Adelson, Noise Removal via Bayesian Wavelet Coring

original With Gaussian noise of std. dev added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR=27.87 (2) Denoised with wavelet marginal model, PSNR=

M. F. Tappen, B. C. Russell, and W. T. Freeman, Efficient graphical models for processing images IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) Washington, DC, 2004.

Motivation for wavelet joint models Note correlations between the amplitudes of each wavelet subband.

Statistics of pairs of wavelet coefficients Contour plots of the joint histogram of various wavelet coefficient pairs Conditional distributions of the corresponding wavelet pairs

(3) Gaussian scale mixtures Wavelet coefficient probability A mixture of Gaussians of scaled covariances observed Gaussian scale mixture model simulation z is a spatially varying hidden variable that can be used to (a) Create the non-gaussian histograms from a mixture of Gaussian densities, and (b) model correlations between the neighboring wavelet coefficients.

original With Gaussian noise of std. dev added, giving PSNR=22.06 (1) Denoised with Gaussian model, PSNR= (3) Denoised with Gaussian scale mixture model, PSNR=30.86 (2) Denoised with wavelet marginal model, PSNR=29.24