Statistical Modeling of Images and its Application into Denoising What is statistics and why? a mathematical science pertaining to the collection, analysis,

Slides:



Advertisements
Similar presentations
State Space Models. Let { x t :t T} and { y t :t T} denote two vector valued time series that satisfy the system of equations: y t = A t x t + v t (The.
Advertisements

ELG5377 Adaptive Signal Processing
Probabilistic Reasoning over Time
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Periodograms Bartlett Windows Data Windowing Blackman-Tukey Resources:
CHAPTER 2 Building Empirical Model. Basic Statistical Concepts Consider this situation: The tension bond strength of portland cement mortar is an important.
Statistics review of basic probability and statistics.
AGC DSP AGC DSP Professor A G Constantinides©1 Modern Spectral Estimation Modern Spectral Estimation is based on a priori assumptions on the manner, the.
The General Linear Model. The Simple Linear Model Linear Regression.
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
STAT 497 APPLIED TIME SERIES ANALYSIS
1 Alberto Montanari University of Bologna Simulation of synthetic series through stochastic processes.
Visual Recognition Tutorial
Variance and covariance M contains the mean Sums of squares General additive models.
The Simple Linear Regression Model: Specification and Estimation
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
EE565 Advanced Image Processing Copyright Xin Li Statistical Modeling of Natural Images in the Wavelet Space Parametric models of wavelet coefficients.
Business Statistics - QBM117 Statistical inference for regression.
Introduction to Regression Analysis, Chapter 13,
Modern Navigation Thomas Herring
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Review of Probability.
1 Patch Complexity, Finite Pixel Correlations and Optimal Denoising Anat Levin, Boaz Nadler, Fredo Durand and Bill Freeman Weizmann Institute, MIT CSAIL.
STAT 497 LECTURE NOTES 2.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Definitions Random Signal Analysis (Review) Discrete Random Signals Random.
Speech Signal Representations I Seminar Speech Recognition 2002 F.R. Verhage.
Data analyses 2008 Lecture Last Lecture Basic statistics Testing Linear regression parameters Skill.
2. Stationary Processes and Models
Why do we Need Statistical Model in the first place? Any image processing algorithm has to work on a collection (class) of images instead of a single one.
Linear Predictive Analysis 主講人:虞台文. Contents Introduction Basic Principles of Linear Predictive Analysis The Autocorrelation Method The Covariance Method.
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
EE565 Advanced Image Processing Copyright Xin Li Image Denoising Theory of linear estimation Spatial domain denoising techniques Conventional Wiener.
CHAPTER 5 SIGNAL SPACE ANALYSIS
EE565 Advanced Image Processing Copyright Xin Li Why do we Need Image Model in the first place? Any image processing algorithm has to work on a collection.
EE565 Advanced Image Processing Copyright Xin Li Image Denoising: a Statistical Approach Linear estimation theory summary Spatial domain denoising.
Robotics Research Laboratory 1 Chapter 7 Multivariable and Optimal Control.
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
ENEE631 Digital Image Processing (Spring'04) Basics on 2-D Random Signal Spring ’04 Instructor: Min Wu ECE Department, Univ. of Maryland, College Park.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: MLLR For Two Gaussians Mean and Variance Adaptation MATLB Example Resources:
Tracking with dynamics
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
EE565 Advanced Image Processing Copyright Xin Li Further Improvements Gaussian scalar mixture (GSM) based denoising* (Portilla et al.’ 2003) Instead.
Basic Random Processes. Introduction Annual summer rainfall in Rhode Island is a physical process has been ongoing for all time and will continue. We’d.
EE565 Advanced Image Processing Copyright Xin Li Why do we Need Image Model in the first place? Any image processing algorithm has to work on a collection.
Random Signals Basic concepts Bibliography Oppenheim’s book, Appendix A. Except A.5. We study a few things that are not in the book.
EC 827 Module 2 Forecasting a Single Variable from its own History.
Chapter 6 Random Processes
Biointelligence Laboratory, Seoul National University
Stochastic Process - Introduction
Linear Regression Modelling
Statistical Estimation
The Chinese University of Hong Kong
6-1 Introduction To Empirical Models
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
STOCHASTIC HYDROLOGY Random Processes
Presentation transcript:

Statistical Modeling of Images and its Application into Denoising What is statistics and why? a mathematical science pertaining to the collection, analysis, interpretation or explanation, and presentation of data What is signal and noise? Jewelry vs. stones (but don’t be fooled by the appearance) What is the risk of statistical approach? Data-driven vs. model-based EE565 Advanced Image Processing Copyright Xin Li

Why do we Need Statistical Model in the first place? Any image processing algorithm has to work on a collection (class) of images instead of a single one Mathematical model gives us the abstraction of common properties of the images within the same class Model is our hypothesis and images are our observation data In physics, can F=ma explain the relationship between force and acceleration?  In image processing, can this model fit this class of images?

Introduction to Statistical Models Motivating applications: Texture synthesis vs. image denoising Statistical image modeling Modeling correlation/dependency Transform-domain texture synthesis Nonparametric texture synthesis Performance evaluation issue

Computer Graphics in SPORE

What is Image/Texture Model? speech Analysis Synthesis Pitch, LPC Residues … texture Analysis Synthesis P(X): parametric /nonparametric

How do we Tell the Goodness of a Model? Synthesis (in statistical language, it is called sampling) Hypothesized model Does the generated sample (experimental result) look like the data of our interests? A fair coin? Does the generated sequence (experimental result) contain the same number of Heads and Tails? Flip the coin Computer simulation

Discrete Random Variables (taken from EE465) Example III: For a gray-scale image (L=256), we can use the notation p(r k ), k = 0,1, …, L - 1, to denote the histogram of an image with L possible gray levels, r k, k = 0,1, …, L - 1, where p(r k ) is the probability of the kth gray level (random event) occurring. The discrete random variables in this case are gray levels. Question: What is wroning with viewing all pixels as being generated from an independent identically distributed (i.i.d.) random variable

To Understand the Problem Theoretically, if all pixels are indeed i.i.d., then random permutation of pixels should produce another image of the same class (natural images) Experimentally, we can write a simple MATLAB function to implement and test the impact of random permutation

Permutated image with identical histogram to lena

Random Process Random process is the foundation for doing research in the field of communication and signal processing (that is why EE513 is the core requirement for qualified exam) Random processes is the vector generalization of (scalar) random variables

Correlation and Dependency (N=2) If the condition holds, then the two random variables are said to be uncorrelated. From our earlier discussion, we know that if x and y are statistically independent, then p(x, y) = p(x)p(y), in which case we write Thus, we see that if two random variables are statistically independent then they are also uncorrelated. The converse of this statement is not true in general.

Covariance of two Random Variables The moment µ 11 is called the covariance of x and y.

Recall: How to Calculate E(XY)? … X Y Empirical solution: Note: When Y=X, we are getting autocorrelation

Stationary Process* TT+K P(X 1,…,X N )=P(X K+1,…,X K+N ) for any K,N (all statistics is time invariant) N N space/time location order of statistics

Gaussian Process With mean vector m and covariance matrix C For convenience, we often assume zero mean (if it is nonzero mean, we can subtract the mean) The question is: is the distribution of observation data Gaussian or not? For Gaussian process, it is stationary as long as its first and second order statistics are time-invariant

The Curse of Dimensionality Even for a small-size image such as 64-by-64, we need to model it by a random process in 4096-dimensional space (R 4096 ) whose covariance matrix is sized by 4096-by-4096 Curse of dimensionality was pointed out by E. Bellman in 1960s; but even computing resource today cannot handle the brute-force search of nearest-neighbor search in relatively high-dimensional space.

Markovian Assumption Andrei A. Markov Pafnuty L. Chebyshev Andrey N. Kolmogorov

A Simple Idea The future is determined by the present but is independent of the past Note that stationarity and Markovianity are two “orthogonal” perspectives of imposing constraints to random processes

Markov Process N-th order Markovian N past samples Parametric or non-parametric characterization

Autoregressive (AR) Model Parametric model (Linear Prediction) An infinite impulse response (IIR) filter z-transform

Example: AR(1) Autocorrelation function a=0.9 k r(k)

Yule-Walker Equation Covariance C

Wiener’s Ideas In practice, we do not know autocorrelation functions but only observation data X 1,…,X M Approach 1: empirically estimate r(k) from X 1,…,X M Approach 2: Formulate the minimization problem of Exercise: you can verify they end up with the same results

Least-Square Estimation M equations, N unknown variables

Least-Square Estimation (Con’d) If you write it out, it is exactly the empirical way of estimating autocorrelation functions – now you have got the third approach R xx rxrx

From 1D to 2D X m,n Causal neighborhood Noncausal neighborhood 678 Causality of neighborhood depends on different applications (e.g., coding vs. synthesis)

Experimental Justifications original Analysis Synthesis random excitation AR model parameters

Failure Example (I) Analysis and Synthesis N=8,M=4096 Another way to look at it: if X and Y are two images of disks, will (X+Y)/2 produce another disk image?

Failure Example (II) Analysis and Synthesis Note that the failure reason of this example is different from the last example (N is not large enough) N=8,M=4096

Summary of AR Modeling Simple and admit closed-form solution Widely studied in time series analysis and speech processing applications Known as 2D Kalman filtering and Gaussian MRF in the literature of image processing Computational issues In 1D scenario, fast algorithms exist due to the Toeplitz property of covariance matrix (e.g., Levinson-Durbin recursion)

Improvement over AR Model Doubly stochastic process* In stationary Gaussian process, second- order statistics are time/spatial invariance In doubly stochastic process, second-order statistics (e.g., covariance) are modeled by another random process with hidden variables Windowing technique To estimate spatially varying statistics

Why do We need Windows? Nothing to do with Microsoft All images have finite dimensions – they can be viewed as the “windowed” version of natural scenes Any empirical estimation of statistical attributes (e.g., mean, variance) is based on the assumption that all N samples observe the same distribution However, how do we know this assumption is satisfied?

1D Rectangular Window X(n) n W=(2T+1)

2D Rectangular Window W=(2T+1) Loosely speaking, parameter estimation from a localized window is a compromised solution to handle spatially varying statistics Such idea is common to other types of non-stationary signals too (e.g., short-time speech processing)

Example As window slides though the image, we will observe that AR model parameters vary from location to location A B C Q: AR coefficients at B and C differ from those at A but for different reasons, Why?