The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

Complex Networks for Representation and Characterization of Images For CS790g Project Bingdong Li 9/23/2009.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Presented by Xinyu Chang
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Chapter 8 Content-Based Image Retrieval. Query By Keyword: Some textual attributes (keywords) should be maintained for each image. The image can be indexed.
Lecture 6: Feature matching CS4670: Computer Vision Noah Snavely.
1 Improving Entropy Registration Theodor D. Richardson.
(1) Feature-point matching by D.J.Duff for CompVis Online: Feature Point Matching Detection,
A Study of Approaches for Object Recognition
1 The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.
Scale Saliency Timor Kadir, Michael Brady Pat Tressel 13-Apr-2005.
(1) Feature-point matching by D.J.Duff for CompVis Online: Feature Point Matching Detection, Extraction.
Object Recognition Using Distinctive Image Feature From Scale-Invariant Key point D. Lowe, IJCV 2004 Presenting – Anat Kaspi.
1 Segmentation of Salient Regions in Outdoor Scenes Using Imagery and 3-D Data Gunhee Kim Daniel Huber Martial Hebert Carnegie Mellon University Robotics.
CS 223B Assignment 1 Help Session Dan Maynes-Aminzade.
Scale Invariant Feature Transform (SIFT)
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Computer vision.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Local invariant features Cordelia Schmid INRIA, Grenoble.
CSE 185 Introduction to Computer Vision Pattern Recognition 2.
Lecture 7: Features Part 2 CS4670/5670: Computer Vision Noah Snavely.
Distinctive Image Features from Scale-Invariant Keypoints Ronnie Bajwa Sameer Pawar * * Adapted from slides found online by Michael Kowalski, Lehigh University.
Distinctive Image Features from Scale-Invariant Keypoints David Lowe Presented by Tony X. Han March 11, 2008.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
CSE 185 Introduction to Computer Vision Feature Matching.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
Presented by David Lee 3/20/2006
776 Computer Vision Jan-Michael Frahm Spring 2012.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
Blob detection.
Machine learning & object recognition Cordelia Schmid Jakob Verbeek.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Another Example: Circle Detection
SIFT Scale-Invariant Feature Transform David Lowe
Interest Points EE/CSE 576 Linda Shapiro.
Presented by David Lee 3/20/2006
Lecture 07 13/12/2011 Shai Avidan הבהרה: החומר המחייב הוא החומר הנלמד בכיתה ולא זה המופיע / לא מופיע במצגת.
Distinctive Image Features from Scale-Invariant Keypoints
Project 1: hybrid images
TP12 - Local features: detection and description
SLAM – Loop Closing with Visually Salient Features
PRESENTED BY Yang Jiao Timo Ahonen, Matti Pietikainen
SURF: Speeded-Up Robust Features
Feature description and matching
Saliency, Scale and Image Description (by T. Kadir and M
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Detecting Artifacts and Textures in Wavelet Coded Images
The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.
CSE 455 – Guest Lectures 3 lectures Contact Interest points 1
CSDD Features: Center-Surround Distribution Distance
From a presentation by Jimmy Huff Modified by Josiah Yoder
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
Histogram Probability distribution of the different grays in an image.
SIFT keypoint detection
SIFT.
The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.
CSE 185 Introduction to Computer Vision
Lecture VI: Corner and Blob Detection
Feature descriptors and matching
The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.
Presented by Xu Miao April 20, 2005
Recognition and Matching based on local invariant features
The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford.
Random Neural Network Texture Model
Presentation transcript:

The Kadir Operator Saliency, Scale and Image Description Timor Kadir and Michael Brady University of Oxford

The issues… salient – standing out from the rest, noticeable, conspicous, prominent scale – find the best scale for a feature image description – create a descriptor for use in object recognition

Early Vision Motivation pre-attentive stage: features pop out attentive stage: relationships between features and grouping

Detection of Salient Features for an Object Class

How do we do this? fixed size windows (simple approach) Harris detector, Lowe detector, etc. Kadir’s approach

Kadir’s Approach Scale is intimately related to the problem of determining saliency and extracting relevant descriptions. Saliency is related to the local image complexity, ie. Shannon entropy. entropy definition H = -∑ Pi log2 Pi i in set of interest

Specifically x is a point on the image Rx is its local neighborhood D is a descriptor and has values {d1, ... dr}. PD,Rx(di) is the probability of descriptor D taking the value di in the local region Rx. (The normalized histogram of the gray tones in a region estimates this probability distribution.)

Local Histograms of Intensity Neighborhoods with structure have flatter distributions which converts to higher entropy.

Problems Kadir wanted to solve Scale should not be a global, preselected parameter Highly textured regions can score high on entropy, but not be useful The algorithm should not be sensitive to small changes in the image or noise.

Kadir’s Methodology use a scale-space approach features will exist over multiple scales Berghoml (1986) regarded features (edges) that existed over multiple scales as best. Kadir took the opposite approach. He considers these too self-similar. Instead he looks for peaks in (weighted) entropy over the scales.

The Algorithm For each pixel location x For each scale s between smin and smax Measure the local descriptor values within a window of scale s Estimate the local PDF (use a histogram) Select scales (set S) for which the entropy is peaked (S may be empty) Weight the entropy values in S by the sum of absolute difference of the PDFs of the local descriptor around S.

Finding salient points the math for saliency discretized saliency entropy weight based on difference between scales s (gray tones) X probability of descriptor D taking value d in the region centered at x with scale s = normalized histogram count for the bin representing gray tone d.

Picking salient points and their scales

Getting rid of texture One goal was to not select highly textured regions such as grass or bushes, which are not the type of objects the Oxford group wanted to recognize Such regions are highly salient with just entropy, because they contain a lot of gray tones in roughly equal proportions But they are similar at different scales and thus the weights make them go away

Salient Regions Instead of just selecting the most salient points (based on weighted entropy), select salient regions (more robust). Regions are like volumes in scale space. Kadir used clustering to group selected points into regions. We found the clustering was a critical step.

More examples

Robustness Claims scale invariant (chooses its scale) rotation invariant (uses circular regions and histograms) somewhat illumination invariant (why?) not affine invariant (able to handle small changes in viewpoint)

More Examples

Temple

Capitol

Houses and Boats

Houses and Boats

Sky Scraper

Car

Trucks

Fish

Other …

Symmetry and More

Benefits General feature: not tied to any specific object Can be used to detect rather complex objects that are not all one color Location invariant, rotation invariant Selects relevant scale, so scale invariant What else is good? Anything bad?

References Kadir, Brady; Saliency, scale and image description; IJCV 45(2), 83-105, 2001 Kadir, Brady; Scale saliency: a novel approach to salient feature and scale selection Treisman; Visual coding of features and objects: Some evidence from behavioral studies; Advances in the Modularity of Vision Selections; NAS Press, 1990 Wolfe, Treisman, Horowitz; What shall we do with the preattentive processing stage: Use it or lose it? (poster); 3rd Annual Mtg Vis Sci Soc

References Dayan, Abbott; Theoretical Neuroscience; MIT Press, 2001 Lamme; Separate neural definitions of visual consciousness and visual attention; Neural Networks 17, 861-872, 2004 Di Lollo, Kawahara, Zuvic, Visser; The preattentive emperor has no clothes: A dynamic redressing; J Experimental Psych, General 130(3), 479-492 Hochstein, Ahissar; View from the top: Hierarchies and reverse hierarchies in the visual system; Neuron 36, 791-804, 2002