Research Methods Lecturer: Steve Maybank

Slides:



Advertisements
Similar presentations
Research Methods Lecturer: Steve Maybank
Advertisements

EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
JPEG Compresses real images Standard set by the Joint Photographic Experts Group in 1991.
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
Chapter 3 Image Enhancement in the Spatial Domain.
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Machine Learning Lecture 8 Data Processing and Representation
Intensity Transformations (Chapter 3)
Digital Image Processing
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Prof. Harriet Fell Fall 2012 Lecture 7 – September 19, 2012
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
School of Computing Science Simon Fraser University
Image processing. Image operations Operations on an image –Linear filtering –Non-linear filtering –Transformations –Noise removal –Segmentation.
L15:Microarray analysis (Classification) The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Dimensional reduction, PCA
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Information Theory and Learning
1 Image Filtering Readings: Ch 5: 5.4, 5.5, 5.6,5.7.3, 5.8 (This lecture does not follow the book.) Images by Pawan SinhaPawan Sinha formal terminology.
Data Basics. Data Matrix Many datasets can be represented as a data matrix. Rows corresponding to entities Columns represents attributes. N: size of the.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
7/2/2015 IENG 486 Statistical Quality & Process Control 1 IENG Lecture 05 Interpreting Variation Using Distributions.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Lecture II-2: Probability Review
Statistical Color Models (SCM) Kyungnam Kim. Contents Introduction Trivariate Gaussian model Chromaticity models –Fixed planar chromaticity models –Zhu.
Separate multivariate observations
Linear Algebra and Image Processing
Pairs of Random Variables Random Process. Introduction  In this lecture you will study:  Joint pmf, cdf, and pdf  Joint moments  The degree of “correlation”
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
R. Ray and K. Chen, department of Computer Science engineering  Abstract The proposed approach is a distortion-specific blind image quality assessment.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
Digital Image Processing Image Compression
An introduction to audio/video compression Dr. Malcolm Wilson.
CSE 185 Introduction to Computer Vision Face Recognition.
23 November Md. Tanvir Al Amin (Presenter) Anupam Bhattacharjee Department of Computer Science and Engineering,
CHAPTER 5 SIGNAL SPACE ANALYSIS
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Class 26: Question 1 1.An orthogonal basis for A 2.An orthogonal basis for the column space of A 3.An orthogonal basis for the row space of A 4.An orthogonal.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
1 Probability and Statistical Inference (9th Edition) Chapter 4 Bivariate Distributions November 4, 2015.
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 11 COMPRESSION.
3D Reconstruction Using Image Sequence
Image Processing Architecture, © Oleh TretiakPage 1Lecture 4 ECE-C490 Winter 2004 Image Processing Architecture Lecture 4, 1/20/2004 Principles.
Independent Component Analysis Independent Component Analysis.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
EE 7730 Image Enhancement. Bahadir K. Gunturk2 Image Enhancement The objective of image enhancement is to process an image so that the result is more.
Feature Extraction 主講人:虞台文.
Chapter 13 Discrete Image Transforms
Independent Component Analysis features of Color & Stereo images Authors: Patrik O. Hoyer Aapo Hyvarinen CIS 526: Neural Computation Presented by: Ajay.
Image Processing Architecture, © Oleh TretiakPage 1Lecture 5 ECEC 453 Image Processing Architecture Lecture 5, 1/22/2004 Rate-Distortion Theory,
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Journal of Vision. 2009;9(3):5. doi: /9.3.5 Figure Legend:
The Chinese University of Hong Kong
Principal Component Analysis (PCA)
Fourier Transform: Real-World Images
ECE 638: Principles of Digital Color Imaging Systems
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
4. DIGITAL IMAGE TRANSFORMS 4.1. Introduction
Image Compression Techniques
Image Coding and Compression
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Research Methods Lecturer: Steve Maybank Department of Computer Science and Information Systems sjmaybank@dcs.bbk.ac.uk Autumn 2015 Data Research Methods in Computer Vision 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Digital Images 95 110 40 34 125 108 25 91 158 116 59 112 166 132 101 124 A digital image is a rectangular array of pixels. Each pixel has a position and a value. Original colour image from the Efficient Content Based Retrieval Group, University of Washington 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Size of Images Digital camera, 5,000x5,000 pixels, 3 bytes/pixel -> 75 MB. Surveillance camera at 25 f/s -> 1875 MB/s. 1000 surveillance cameras -> ~1.9 TB/s. Not all of these images are useful! 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Image Compression Divide the image into blocks, and compress each block separately, e.g. JPEG uses 8x8 blocks. Lossfree compression: the original image can be recovered exactly from the compressed image. Lossy compression: the original image cannot be recovered. 11 November 2015 Birkbeck College, U. London

Why is Compression Possible? Natural image: values of neighbouring pixels are strongly correlated. White noise image: values of neighbouring pixels are not correlated. Compression discards information. 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Measurement Space Vectors from 8x8 blocks R64 Each 8x8 block yields a vector in R64. The vectors from natural images tend to lie in a low dimensional subspace of R64. 11 November 2015 Birkbeck College, U. London

Strategy for Compression vectors from 8x8 blocks Choose a basis for R64 in which the low dimensional subspace is spanned by the first few coordinate vectors. Retain these coordinates and discard the rest. 11 November 2015 Birkbeck College, U. London

Discrete Cosine Transform 11 November 2015 Birkbeck College, U. London

Basis Images for the DCT UTe(1) UTe(2) UTe(3) UTe(4) 11 November 2015 Birkbeck College, U. London

Example of Compression using DCT Image constructed from 3 DCT coefficients in each 8x8 block. Original image 11 November 2015 Birkbeck College, U. London

Linear Classification y y Given two sets X, Y of measurement vectors from different classes, find a hyperplane that separates X and Y. y y x x x x x x A new vector is assigned to the class of X or to the class of Y, depending on its position relative to the hyperplane. 11 November 2015 Birkbeck College, U. London

Fisher Linear Discriminant 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Maximisation of J(v) 11 November 2015 Birkbeck College, U. London

Edge Regions and Regions Without a Peak 3x3 blocks such that the grey level of the central pixel equals the mean grey level of the 9 pixels 3x3 blocks with large Sobel gradients 11 November 2015 Birkbeck College, U. London

Projections Onto a Random Plane Sobel edges Regions without a peak Superposed plots 11 November 2015 Birkbeck College, U. London

Projections Onto a 1-Dimensional FLD Sobel edges Regions without a peak Combined histograms 11 November 2015 Birkbeck College, U. London

Discrete Distribution A probability distribution on a discrete set S={1, 2,…, n} is a set of numbers 𝑝 𝑖 such that 0≤ 𝑝 𝑖 ≤1 𝑖=1 𝑛 𝑝 𝑖 =1 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Interpretations Bayes: 𝑝 𝑖 is a measure of our knowledge that item i will be chosen from S. Frequentist: if a large number m of independent samples from S are chosen then i occurs approximately 𝑚 𝑝 𝑖 times 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Terminology Event: subset of S Probability of event E: P E = 𝑖𝜖𝐸 𝑝 𝑖 Conditional Probability: 𝑃 𝐸 𝐹 = 𝑃(𝐸∩𝐹) 𝑃(𝐹) 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Example Roll two dice. F=event that total is 8. S={(i, j), 1<=i, j<=6} The pairs (i, j) all have the same probability, thus P({i, j})=1/36, 1<=i<=36 F={(6,2), (2,6), (3,5), (5,3), (4,4)} P(F) = 5/36 11 November 2015 Birkbeck College, U. London

Example of a Conditional Probability E={(6,2)}. What is the probability of E given F? 𝑃 𝐸 𝐹 = 𝑃 𝐸∩𝐹 𝑃 𝐹 = 𝑃 𝐸 𝑃 𝐹 = 1 36 5/36 =1/5 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Independent Events The events E, F are independent if 𝑃 𝐸∩𝐹 =𝑃 𝐸 𝑃 𝐹 Example: E=first number is 6 F=second number is 5 𝑃 𝐸∩𝐹 = 1 36 P(E) = 1/6, P(F)=1/6 11 November 2015 Birkbeck College, U. London

Probability Density Function A pdf of the real line R is a function 𝑓:𝑅→𝑅 such that 𝑓 𝑥 ≥0,𝑥𝜖𝑅 −∞ ∞ 𝑓 𝑥 𝑑𝑥=1 A pdf is used to assign probabilities to subsets of R: P(A) = 𝐴 𝑓𝑑𝑥 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London The Gaussian PDF 𝑓 𝑥 = (2𝜋) −1/2 𝑒 − 𝑥 2 /2 Mean value: −∞ ∞ 𝑥𝑓 𝑥 𝑑𝑥 Variance: −∞ ∞ 𝑥 2 𝑓 𝑥 𝑑𝑥 11 November 2011 Birkbeck College, U. London

Estimation of Parameters Given samples 𝑥 1 , 𝑥 2 , … 𝑥 𝑛 in R from a probability distribution, estimate the pdf, assuming it is Gaussian Mean value: 𝜇= 1 𝑛 𝑖=1 𝑛 𝑥 𝑖 Variance: 𝜎 2 = 1 𝑛 𝑖=1 𝑛 ( 𝑥 𝑖 −𝜇) 2 𝑓 𝑥 =(2𝜋 𝜎 2 ) −1/2 𝑒 − 𝑥−𝜇 2 /(2 𝜎 2 ) 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Gaussian pdf in 2D 𝑓 𝑥,𝑦 = (2𝜋) −1 𝑒 −( 𝑥 2 + 𝑦 2 )/2 11 November 2015 Birkbeck College, U. London

Estimation of a Multivariate Gaussian Mean value in 𝑅 𝑑 : μ= 1 𝑛 𝑖=1 𝑛 𝑥 𝑖 Covariance matrix: 𝐶 𝑗𝑘 = 1 𝑛 𝑖=1 𝑛 ( 𝑥 𝑖 −𝜇) 𝑗 ( 𝑥 𝑖 −𝜇) 𝑘 𝑓 𝑥 = 2𝜋 det 𝐶 − 𝑑 2 𝐸𝑥𝑝 − 1 2 𝑥−𝜇 𝑇 𝐶 −1 (𝑥−𝜇) 11 November 2015 Birkbeck College, U. London

Histogram of a DCT Coefficient The pdf is leptokurtic, i.e. it has a peak at 0 and “fat tails”. 11 November 2015 Birkbeck College, U. London

Sparseness of the DCT Coefficients For a given 8×8 block, only a few DCT coefficients are significantly different from 0. Given a DCT coefficient of low to moderate frequency, there exist some blocks for which it is large. 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Whitening the Data 11 November 2015 Birkbeck College, U. London

Independence and Correlation 11 November 2015 Birkbeck College, U. London

Independent Components Analysis Let z be a rv with values in R^n such that cov(z)=I. Find an orthogonal matrix V such that the components of s =Vz are as independent as possible. 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London PDF for Image Blocks Assume a particular leptokurtic pdf p for the components si, e.g. a two sided Laplace density. Find the orthogonal matrix V with rows v1,…,vn that maximises the product p(s1)…p(sn) = p(v1.z)…p(vn.z) 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Example Natural Image Statistics: a probabilistic approach to early computational vision, by A. Hyvarinen, J. Hurri and P.O. Hoyer. Page 161. ICA applied to image patches yields Gabor type filters similar to those found in the human visual system. 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Stereo Pair of Images UC Berkeley looking west http://arch.ced.berkeley.edu/kap/wind/?p=32 11 November 2015 Birkbeck College, U. London

Birkbeck College, U. London Salience A left hand image region is salient if the correct right hand matching region can be found with a high probability. H: hypothesis that (v(1),v(2)) is a matching pair. B: hypothesis that (v(1),v(2)) is a background pair 11 November 2015 Birkbeck College, U. London

Formulation using PDFs The region yielding v(1) is salient if p(v(2)|v(1), H) differs significantly from p(v(2)|v(1), B) Measurement of difference: the Kullback-Leibler divergence: 11 November 2015 Birkbeck College, U. London

Illustration of the Kullback-Leibler Divergence KL divergence = 0.125 KL divergence = 0.659 See S.J. Maybank (2013) A probabilistic definition of salient regions for image matching. Neurocomputing v. 120, pp. 4-14. 11 November 2015 Birkbeck College, U. London