A Model of Saliency-Based Visual Attention for Rapid Scene Analysis

Slides:



Advertisements
Similar presentations
CSCE 643 Computer Vision: Template Matching, Image Pyramids and Denoising Jinxiang Chai.
Advertisements

A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.
Boundary Detection - Edges Boundaries of objects –Usually different materials/orientations, intensity changes.
Hierarchical Saliency Detection School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
黃文中 Preview 2 3 The Saliency Map is a topographically arranged map that represents visual saliency of a corresponding visual scene. 4.
Templates and Image Pyramids Computational Photography Derek Hoiem, University of Illinois 09/06/11.
A Novel Method for Generation of Motion Saliency Yang Xia, Ruimin Hu, Zhenkun Huang, and Yin Su ICIP 2010.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
CSCE 641 Computer Graphics: Image Sampling and Reconstruction Jinxiang Chai.
CSCE 641 Computer Graphics: Image Sampling and Reconstruction Jinxiang Chai.
Visual Attention More information in visual field than we can process at a given moment Solutions Shifts of Visual Attention related to eye movements Some.
Computer Vision Introduction to Image formats, reading and writing images, and image environments Image filtering.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Visual Attention Jeremy Wyatt. Where to look? Many visual processes are expensive Humans don’t process the whole visual field How do we decide what to.
Christian Siagian Laurent Itti Univ. Southern California, CA, USA
LYU 0602 Automatic PhotoHunt Generation Automatic PhotoHunt Generation Shum Hei Lung To Wan Chi Supervisor: Prof. Michael R. Lyu.
Michigan State University 1 “Saliency-Based Visual Attention” “Computational Modeling of Visual Attention”, Itti, Koch, (Nature Reviews – Neuroscience.
Lecture 2: Image filtering
Attention in Computer Vision Mica Arie-Nachimson and Michal Kiwkowitz May 22, 2005 Advanced Topics in Computer Vision Weizmann Institute of Science.
Introduction of Saliency Map
ENG4BF3 Medical Image Processing
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30 – 11:50am.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
CS559: Computer Graphics Lecture 3: Digital Image Representation Li Zhang Spring 2008.
CSNDSP’06 Visual Attention based Region of Interest Coding for Video - telephony Applications Nicolas Tsapatsoulis Computer Science Dept. University of.
1 © 2010 Cengage Learning Engineering. All Rights Reserved. 1 Introduction to Digital Image Processing with MATLAB ® Asia Edition McAndrew ‧ Wang ‧ Tseng.
Text Lecture 2 Schwartz.  The problem for the designer is to ensure all visual queries can be effectively and rapidly served.  Semantically meaningful.
報告人:張景舜 P.H. Wu, C.C. Chen, J.J. Ding, C.Y. Hsu, and Y.W. Huang IEEE Transactions on Image Processing, Vol. 22, No. 9, September 2013 Salient Region Detection.
黃文中 Introduction The Model Results Conclusion 2.
Templates and Image Pyramids Computational Photography Derek Hoiem, University of Illinois 09/03/15.
Instructor: S. Narasimhan
1 Computational Vision CSCI 363, Fall 2012 Lecture 5 The Retina.
Department of Psychology & The Human Computer Interaction Program Vision Sciences Society’s Annual Meeting, Sarasota, FL May 13, 2007 Jeremiah D. Still,
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
Creating Better Thumbnails Chris Waclawik. Project Motivation Thumbnails used to quickly select a specific a specific image from a set (when lacking appropriate.
Templates, Image Pyramids, and Filter Banks
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
1 Perception and VR MONT 104S, Fall 2008 Lecture 4 Lightness, Brightness and Edges.
Autonomous Robots Vision © Manfred Huber 2014.
Street Smarts: Visual Attention on the Go Alexander Patrikalakis May 13, XXX.
Efficient Color Boundary Detection with Color-opponent Mechanisms CVPR2013 Posters.
Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos VC 15/16 – TP7 Spatial Filters Miguel Tavares Coimbra.
1 Computational Vision CSCI 363, Fall 2012 Lecture 6 Edge Detection.
1 Computational Vision CSCI 363, Fall 2012 Lecture 16 Stereopsis.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Computer vision. Applications and Algorithms in CV Tutorial 3: Multi scale signal representation Pyramids DFT - Discrete Fourier transform.
Instructor: Mircea Nicolescu Lecture 5 CS 485 / 685 Computer Vision.
Last Lecture photomatix.com. Today Image Processing: from basic concepts to latest techniques Filtering Edge detection Re-sampling and aliasing Image.
 Mentor : Prof. Amitabha Mukerjee Learning to Detect Salient Objects Team Members - Avinash Koyya Diwakar Chauhan.
Miguel Tavares Coimbra
Edge Detection slides taken and adapted from public websites:
Image Pyramids and Applications
Implementation of a Visual Attention Model
Lecture 2: Edge detection
Early Processing in Biological Vision
Jeremy Bolton, PhD Assistant Teaching Professor
Lecture 2: Edge detection
Volume 55, Issue 3, Pages (August 2007)
Motion-Based Analysis of Spatial Patterns by the Human Visual System
Edge Detection Today’s reading
Edge Detection in Computer Vision
Edge Detection Today’s readings Cipolla and Gee Watt,
Coding of Natural Scenes in Primary Visual Cortex
Origin and Function of Tuning Diversity in Macaque Visual Cortex
Volume 64, Issue 6, Pages (December 2009)
Lecture 2: Edge detection
Valerio Mante, Vincent Bonin, Matteo Carandini  Neuron 
Maxwell H. Turner, Fred Rieke  Neuron 
Presentation transcript:

A Model of Saliency-Based Visual Attention for Rapid Scene Analysis By: Laurent Itti, Christof Koch, and Ernst Niebur IEEE TRANSACTIONS, NOVEMBER 1998 Presenter: Vahid Vahidi Fall 2015

Outline Definition and the goal of this paper Method Linear filtering Center surround differences and normalization Across scale combinations and normalization Linear combination and Winner-take-all Simulation results Comparison with Spatial Frequency Content Models Comparing white noise with colored noise Strengths and Limitations Summary

Definition and the goal of this paper Which part of the image attracts more attention? Goal Simulation of what is going on in our brain

Method

Linear filtering Four broadly-tuned color channels are created: R = r - (g + b)/2 for red, G = g - (r + b)/2 for green, B = b - (r + g)/2 for blue, and Y = (r + g)/2 - |r - g|/2 - b for yellow With r, g, and b being the red, green, and blue channels of the input image, an intensity image I is obtained as I=(r+g+b)/3. Nine spatial scales are created using dyadic Gaussian pyramids. Low-pass filter and subsample the input image would be performed. 1:1 (scale zero) to 1:256 (scale eight) in eight octaves.

Center surround differences and normalization Compute center-surround differences to determine contrast, by taking the difference between a fine (center) and a coarse scale (surround) for a given feature. This operation across spatial scales is done by interpolation to the fine scale and then point-by-point subtraction.

Normalization Map normalization operator: Promotes maps in which a small number of strong peaks of activity is present Suppressing maps which contain numerous comparable peak responses.

Across scale combinations and normalization The feature maps are combined into three conspicuity maps at the scale 4. This is obtained through across-scale addition by reducing each map to scale 4 and point-by-point addition.

Linear combination and Winner-take-all saliency map would be achieved Winner-take-all Models the SM as a 2D layer of leaky integrate-and- fire neurons at scale four. These model neurons consist of a single capacitance which integrates the charge When the threshold is reached, a prototypical spike is generated, and the capacitive charge is shunted to zero All WTA neurons also evolve independently of each other, until one (the “winner”) first reaches threshold and fires.

Comparison with Spatial Frequency Content Models SFC At a given image location, a 16 ´ 16 image patch is extracted from each I(2), R(2), G(2), B(2), and Y(2) map, and 2D Fast Fourier Transforms (FFTs) are applied to the patches. The SFC measure is the average of the numbers of non-negligible coefficients in the five corresponding patches.

Comparing white noise with colored noise

Strengths and Limitations In this approach, architecture and components mimic the properties of primate early vision It is capable of strong performance with complex natural scenes (Ex. it quickly detected salient traffic signs) The major strength of this approach lies in the massively parallel implementation Limitations it will fail at detecting unimplemented feature types (e.g., T junctions or line terminators) Without modifying the preattentive feature-extraction stages, our model cannot detect conjunctions of features.

MATLAB MATLAB has saliency Toolbox

Summary I have reviewed the saliency paper Definition of saliency Method Based on intensity, color and orientation, 42 center surround maps would be achieved. Normalized maps are combined at scale 4 Saliency map is achieved by the combination of normalized maps of intensity, color and orientation Winner-take-all procedure finds salient areas in the decreasing order Results Saliency performs better than SFC in presence of noise Saliency performs better in the presence of white noise in comparison to the presence of colored noise

Thanks for your attention