Lecture 2 Math & Probability Background ch. 1-2 of Machine Vision by Wesley E. Snyder & Hairong Qi Spring 2014 BioE 2630 (Pitt) : 16-725 (CMU RI) 18-791.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

3D Geometry for Computer Graphics
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
ECE 472/572 - Digital Image Processing Lecture 8 - Image Restoration – Linear, Position-Invariant Degradations 10/10/11.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Principal Component Analysis
Function Optimization Newton’s Method. Conjugate Gradients
Computer Graphics Recitation 5.
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Motion Analysis (contd.) Slides are from RPI Registration Class.
CSci 6971: Image Registration Lecture 4: First Examples January 23, 2004 Prof. Chuck Stewart, RPI Dr. Luis Ibanez, Kitware Prof. Chuck Stewart, RPI Dr.
Face Recognition Using Embedded Hidden Markov Model.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Image Filtering Readings: Ch 5: 5.4, 5.5, 5.6,5.7.3, 5.8 (This lecture does not follow the book.) Images by Pawan SinhaPawan Sinha formal terminology.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
CS Pattern Recognition Review of Prerequisites in Math and Statistics Prepared by Li Yang Based on Appendix chapters of Pattern Recognition, 4.
Modern Navigation Thomas Herring
Quantum One: Lecture 8. Continuously Indexed Basis Sets.
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka Virginia de Sa (UCSD) Cogsci 108F Linear.
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Linear Algebra and Image Processing
Image Representation Gaussian pyramids Laplacian Pyramids
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Linear Algebra Review 1 CS479/679 Pattern Recognition Dr. George Bebis.
Recap of linear algebra: vectors, matrices, transformations, … Background knowledge for 3DM Marc van Kreveld.
The content of these slides by John Galeotti, © Carnegie Mellon University (CMU), was made possible in part by NIH NLM contract# HHSN P,
+ Review of Linear Algebra Optimization 1/14/10 Recitation Sivaraman Balakrishnan.
 1  Outline  stages and topics in simulation  generation of random variates.
Digital Image Processing, 3rd ed. © 1992–2008 R. C. Gonzalez & R. E. Woods Gonzalez & Woods Matrices and Vectors Objective.
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
Chapter 5 Eigenvalues and Eigenvectors 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
Linear algebra: matrix Eigen-value Problems Eng. Hassan S. Migdadi Part 1.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Motivation Thus far we have dealt primarily with the input/output characteristics of linear systems. State variable, or state space, representations describe.
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
1. Systems of Linear Equations and Matrices (8 Lectures) 1.1 Introduction to Systems of Linear Equations 1.2 Gaussian Elimination 1.3 Matrices and Matrix.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
CS654: Digital Image Analysis Lecture 11: Image Transforms.
Matrices CHAPTER 8.9 ~ Ch _2 Contents  8.9 Power of Matrices 8.9 Power of Matrices  8.10 Orthogonal Matrices 8.10 Orthogonal Matrices 
Unsupervised Learning II Feature Extraction
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Computational Physics (Lecture 7) PHY4061. Eigen Value Problems.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
EEE 431 Computational Methods in Electrodynamics
Lecture 16: Image alignment
Biointelligence Laboratory, Seoul National University
References Jain (a text book?; IP per se; available)
Eigenfaces (for Face Recognition)
Chapter 7. Classification and Prediction
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Review of Linear Algebra
CS479/679 Pattern Recognition Dr. George Bebis
The Chinese University of Hong Kong
Lecture 03: Linear Algebra
Image Restoration and Denoising
2D transformations (a.k.a. warping)
CONTEXT DEPENDENT CLASSIFICATION
Lecture 3 Math & Probability Background ch
Lecture 3 Math & Probability Background ch
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
(CMU ECE) : (CMU BME) : BioE 2630 (Pitt)
Presentation transcript:

Lecture 2 Math & Probability Background ch. 1-2 of Machine Vision by Wesley E. Snyder & Hairong Qi Spring 2014 BioE 2630 (Pitt) : (CMU RI) (CMU ECE) : (CMU BME) Dr. John Galeotti & Prahlad G Menon

General notes about the book The book is an overview of many concepts Top quality design requires: – Reading the cited literature – Reading more literature – Experimentation & validation 2

Two themes in Text Book Consistency – A conceptual tool implemented in many/most algorithms – Often must fuse information from local measurements to make global conclusions about the image – Is the fused input information conflicting..? Optimization – Mathematical mechanism to solve matrix equations. – The “workhorse” of machine vision 3

Image Processing Topics Enhancement Coding – Compression Restoration – “Fix” an image – Requires model of image degradation Reconstruction 4

Machine Vision Topics AKA: – Computer vision – Image analysis – Image understanding Pattern recognition: 1.Measurement of features Features characterize the image, or some part of it 2.Pattern classification Requires knowledge about the possible classes 5 Our Focus Feature Extraction Classification & Further Analysis Original Image

Feature measurement 6 Noise removal Segmentation Original Image Shape AnalysisConsistency Analysis Matching Features Restoration Ch. 6-7 Ch. 8 Ch Ch Ch. 9

Linear Operations What is a linear operation..? How to represent it as a Matrix operator..? Rules for basic matrix multiplication. Matrix inverse & unitary matrix. What is linear independence..? What is a basis..? What is an orthonormal basis set..? Why is it useful..? What is function minimization..? Why is it useful..? Examples of useful linear operations / operators. What is an inverse linear transform..? Example. 7

Linear algebra Unit vector: |x| = 1 Orthogonal vectors: x T y = 0 Orthogonal square matrices: A T = A -1 Orthonormal: orthogonal unit vectors Inner product of continuous functions – Orthogonality & orthonormality apply here too 8

Linear transforms For a matrix, A Quadratic form: Positive definite: – Applies to A if 9

Linear independence No one vector is a linear combination of the others – x j   a i x i for any a i across all i  j Any linearly independent set of d vectors {x i=1…d } is a basis set that spans the space  d – Any other vector in  d may be written as a linear combination of {x i } Often convenient to use orthonormal basis sets Projection: if y=  a i x i then a i =y T x i 10

Vector norms 11

Examples of linear operators Rotation, translation, scaling – spatial transforms Convolution Wolfram – definition of a linear operator / transformation: 12

Simple example of convolution of input image (matrix, S) and impulse response (kernel) in 2D spatial. Y[n] = (S * h) [n]. Notice that the kernel matrix is flipped both horizontal and vertical direction before multiplying the overlapped input data. The Convolution Operator – Example… 13

Eigen values & Eigen vectors 14 Let A be an n x n matrix. The number is an eigenvalue of A if there exists a non-zero vector v such that Av = λ v (The Characteristic equation) In this case, vector v is called an eigen-vector of A corresponding to λ. What does this mean physically..? A maps v onto itself with only a change in length.

Derivatives Of a scalar function of x : – Called the gradient – Really important! Of a vector function of x – Called the Jacobian Hessian = matrix of 2nd derivatives of a scalar function 15

Linear INVERSE problems – de-blurring an image 16 Given measured data, g, and information about the system H and noise model, recover an estimate ~f that closely approximates f. When H is square, solve g = Hf via f = H -1 g directly (exactly) via Fourier Transform methods. Eg: Lucy–Richardson deconvolution, is an iterative procedure for recovering a latent image that has been blurred by a known point spread function.deconvolutioniterative procedurelatent imageblurredpoint spread function

Least Squares Fitting – solving as a linear system (assignment) 17 What kind of basis is ideal for A, here..?

Function minimization Find the vector x which produces a minimum of some function f (x) – x is a parameter vector – f(x) is a scalar function of x The “objective function” The minimum value of f is denoted: The minimizing value of x is denoted: 18

Other Numerical minimization techniques Gradient descent – The derivative points away from the minimum – Take small steps, each one in the “down-hill” direction Local vs. global minima Combinatorial optimization: – Use simulated annealing Image optimization: – Use mean field annealing 19

Fourier Orthonormal Basis Many imaging problems easier to understand if we choose building blocks (k) carefully. = Σ n Φ* k [n] Φ l [n] = 0 if k != l, 20 (1/N) Ac = s, is simply c = A*s: The coefficients c[k] of are the Discrete Fourier Transform of the data s[n].

Orthonormal Expansion Relate the problem of expanding these approximations to solving Ax = b. Now x are the expansion coefficients, b is an image, and A = [Φ1,Φ2, … ], where each Φi is a column vector representing an orthonormal basis. When columns of A are orthonormal: (A*)(A) = I, so x = A T b. b[p] = Σ k x k A p,k, or, s[p] = Σ k c k Φ k [p], where c k may be termed Fourier cofficients, if Φ k [p] refers to the Fourier basis. k[n] = e -j2kn/N = cos(2kn/N) + j sin(2kn/N). = N for Fourier. 21

Many useful properties of Fourier basis… Interesting fact: k-space  MRI data is acquired in frequency space (or Fourier space) and then converted back to image space. 22

Probability Probability of an event a occurring: – Pr(a) Independence – Pr(a) does not depend on the outcome of event b, and vice-versa Joint probability – Pr(a,b) = Prob. of both a and b occurring Conditional probability – Pr(a|b) = Prob. of a if we already know the outcome of event b – Read “probability of a given b ” 23

Probability for continuously-valued functions Probability distribution function: P(x) = Pr(z<x) Probability density function: Expected value 24

Markov models For temporal processes: – The probability of something happening is dependent on a thing that just recently happened. For spatial processes – The probability of something being in a certain state is dependent on the state of something nearby. – Example: The value of a pixel is dependent on the values of its neighboring pixels. 25

Markov chain Simplest Markov process – automatic random transition between finite states with determinable: Example: symbols transmitted one at a time – What is the probability that the next symbol will be w ? For a Markov chain: – “The probability conditioned on all of history is identical to the probability conditioned on the last symbol received.” 26 – Initial probabilities – Transition probabilities

Hidden Markov models (HMMs) 27

HMM switching Governed by a finite state machine (FSM) 28 Output 1 st Process Output 2 nd Process HMM generates observation sequence

The HMM Task Given only the output f (t), determine: 1.The most likely state sequence of the switching FSM – Use the Viterbi algorithm – Computational complexity = (# state changes) * (# state values) 2 – Much better than brute force, which = (# state values) (# state changes) 2.The parameters of each hidden Markov model – Use the iterative process in the book – Better, use someone else’s debugged code that they’ve shared 29

Hidden Markov models (HMMs) 30 1 st Markov Process 2 nd Markov Process f ( t )

Markov models for Image Segmentation 31 czhu/papers/DDMCMC_re print.pdf Code: Download_ddmcmc.htm