1. A given pattern p is sought in an image. The pattern may appear at any location in the image. The image may be subject to some tone changes. 2 pattern.

Slides:



Advertisements
Similar presentations
Image Registration  Mapping of Evolution. Registration Goals Assume the correspondences are known Find such f() and g() such that the images are best.
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Hongliang Li, Senior Member, IEEE, Linfeng Xu, Member, IEEE, and Guanghui Liu Face Hallucination via Similarity Constraints.
Face Alignment with Part-Based Modeling
TP14 - Local features: detection and description Computer Vision, FCUP, 2014 Miguel Coimbra Slides by Prof. Kristen Grauman.
1 Video Processing Lecture on the image part (8+9) Automatic Perception Volker Krüger Aalborg Media Lab Aalborg University Copenhagen
E.G.M. PetrakisFiltering1 Linear Systems Many image processing (filtering) operations are modeled as a linear system Linear System δ(x,y) h(x,y)
Face Recognition and Biometric Systems
Forward-Backward Correlation for Template-Based Tracking Xiao Wang ECE Dept. Clemson University.
Robust Object Tracking via Sparsity-based Collaborative Model
Different Tracking Techniques  1.Gaussian Mixture Model:  1.Construct the model of the Background.  2.Given sequence of background images find the.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
1 Formation et Analyse d’Images Session 6 Daniela Hall 18 November 2004.
1 Formation et Analyse d’Images Session 12 Daniela Hall 16 January 2006.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
Robust and large-scale alignment Image from
Patch Descriptors CSE P 576 Larry Zitnick
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Principal Component Analysis
A Study of Approaches for Object Recognition
Computing motion between images
Feature tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on good features.
Illumination Normalization with Time-Dependent Intrinsic Images for Video Surveillance Yasuyuki Matsushita, Member, IEEE, Ko Nishino, Member, IEEE, Katsushi.
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Shadow Detection In Video Submitted by: Hisham Abu saleh.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Computer vision.
Summarized by Soo-Jin Kim
Image Processing in Freq. Domain Restoration / Enhancement Inverse Filtering Match Filtering / Pattern Detection Tomography.
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Engineering Statistics ENGR 592 Prepared by: Mariam El-Maghraby Date: 26/05/04 Design of Experiments Plackett-Burman Box-Behnken.
Ch 4. Linear Models for Classification (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized and revised by Hee-Woong Lim.
Data Extraction using Image Similarity CIS 601 Image Processing Ajay Kumar Yadav.
Parameter estimation. 2D homography Given a set of (x i,x i ’), compute H (x i ’=Hx i ) 3D to 2D camera projection Given a set of (X i,x i ), compute.
Image Segmentation and Edge Detection Digital Image Processing Instructor: Dr. Cheng-Chien LiuCheng-Chien Liu Department of Earth Sciences National Cheng.
Conclusions The success rate of proposed method is higher than that of traditional MI MI based on GVFI is robust to noise GVFI based on f1 performs better.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
Edges.
CSC2515: Lecture 7 (post) Independent Components Analysis, and Autoencoders Geoffrey Hinton.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Nonlinear Dimension Reduction: Semi-Definite Embedding vs. Local Linear Embedding Li Zhang and Lin Liao.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Lecture 16: Image alignment
Visual homing using PCA-SIFT
A 2 veto for Continuous Wave Searches
CS4670 / 5670: Computer Vision Kavita Bala Lec 27: Stereo.
TP12 - Local features: detection and description
PRESENTED BY Yang Jiao Timo Ahonen, Matti Pietikainen
Fast Preprocessing for Robust Face Sketch Synthesis
Paper Presentation: Shape and Matching
Feature description and matching
The General Linear Model (GLM): the marriage between linear systems and stats FFA.
Fast and Robust Object Tracking with Adaptive Detection
Multi-modality image registration using mutual information based on gradient vector flow Yujun Guo May 1,2006.
IMAGE BASED VISUAL SERVOING
Learning with information of features
Parallelization of Sparse Coding & Dictionary Learning
Presented by: Chang Jia As for: Pattern Recognition
Image Registration 박성진.
CSSE463: Image Recognition Day 30
Feature Detection .
Feature descriptors and matching
CSSE463: Image Recognition Day 30
Image Registration  Mapping of Evolution
Presentation transcript:

1

A given pattern p is sought in an image. The pattern may appear at any location in the image. The image may be subject to some tone changes. 2 pattern image similarity

Source of Tone Changes: –Illumination conditions –Camera parameters –Different Modalities Applications: “patch based” methods –Image summarization –Image retargeting –Image editing –Super resolution –Tracking, Recognition, many more …

Assumption: Tone changes can be locally represented as a Tone Mapping between the sought pattern p and a candidate window w: 4 w=M(p) or p=M(w) VpVp VwVw VpVp

Given a pattern p and a candidate window w a distance metric must be defined, according to which matchings are determined: Desired properties of D(p,w) : –Discriminative –Robust to Noise –Invariant to some deformations: tone mapping –Fast to execute D(p,w)D(p,w)

Sum of Squared Difference (SSD): –By far the most common solution. –Assumes the identity tone mapping. –Fast implementation (~1 convolution).

Normalized Cross Correlation (NCC): –Compensates for linear mappings. –Fast implementation (~ 1 convolution).

Ordinal Coding (LBP, SURF, BRIEF): –Each pixel is assigned a value representing its surrounding structural content. –Compensates for monotonic mappings. –Fast implementation. –Sensitive to noise.

Mutual Information (MI): –Measures the statistical dependencies between p and w. –Compensates for non-linear mappings. H(w) H(p) D(p,w)=I(p,w)=H(w)-H(w | p)

Properties: Measures the entropy loss in w given p. Highly Discriminative. Sensitive to bin-size / kernel-variance. Inaccurate for small patched. Very slow to apply. H(w) H(p) I(w,p)

Properties: Highly discriminative. Invariant to any (non-linear) tone mapping. Robust to noise. Very fast to apply. Natural generalization of the NCC for non-linear mappings.

Proposed distance measure: Note: the division by var(w) avoids the trivial mapping.

Basic Ideas: Approximate M(p) by a piece-wise constant mapping. Enables to represent M(p) in a linear form. The minimization can be solved in closed form. VpVp VwVw VpVp VwVw

Assume the pattern/window values are restricted to the half open interval [a,b). We divide [a,b) into k bins  =[  1,  2,...,  k+1 ] A value z is naturally associated with a single bin: B(z)=j if z  [  j,  j+1 ) z 11 22  k+1 jj  j+1

We define a pattern slice

2 nd slice p 2 1 st slice p We define a pattern slice

Raster scanning the slice windows and stacking into a matrix constructs a slice matrix S p =[p 1 p 2 … p k ]. = S p

*  The matrix S p is orthogonal: p i  p j = |p i |  ij Its columns span the space of piecewise constant tone mappings of p: S p   p

 M(p) *  Changing the  values to a different vector, , applies piece- wise tone mapping:  p S p  S p   M(p)

Representing tone mapping in a linear form, D(p,w) boils down to: The above minimization can be solved in closed form:

 D (, )= ( )  -( )   2 2 p w  2

* * 2 box filter ( ) 2 Pattern

p 2 : ( ) 2 Pattern *

Convolutions can be applied efficiently since p j is sparse. Convolving with p j requires |p j | operations. Since p i  p j =  run time for all k sparse convolutions sum up to a single convolution!

The MTM can be shown to measure: This is related to the Correlation-Ratio distance measure (Pearson 1930, Roche et al. 1998) and the Fisher’s LD. Restricting M to be a linear tone mapping: M(z)=az+b, the MTM distance D(w,p) reduces to the NCC.

MTM and MI are similar in spirit: While MI maximizes the entropy reduction in (w | p) MTM maximizes the variance reduction in (w | p). However, MTM outperforms MI with respect to speed and accuracy (in small patch cases).

28 Detection rates (over 2000 image pattern pairs) v.s extremity of the applied tone mapping.

29 Run time for various pattern sizes (in 512x512 image)

30 Performance of MI and MTM for various pattern sizes and over various bin-sizes

How can we distinguish between target and background? Background model Current video frame

Naïve Background Subtraction

Background model Video frame Assumption: Shadow patches are functionally dependent on the background patches.

MTM distance

A new distance measure that accounts for non- linear tone mappings. A very efficient scheme which can be applied over the entire image using ~1 convolution. A natural generalization of NCC. 36

37