Michigan State University 1 “Saliency-Based Visual Attention” “Computational Modeling of Visual Attention”, Itti, Koch, (Nature Reviews – Neuroscience.

Slides:



Advertisements
Similar presentations
Attention and neglect.
Advertisements

A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.
Object Recognition with Features Inspired by Visual Cortex T. Serre, L. Wolf, T. Poggio Presented by Andrew C. Gallagher Jan. 25, 2007.
HMAX Models Architecture Jim Mutch March 31, 2010.
Attention - Overview Definition Theories of Attention Neural Correlates of Attention Human neurophysiology and neuroimaging Change Blindness Deficits of.
Hierarchical Saliency Detection School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
黃文中 Preview 2 3 The Saliency Map is a topographically arranged map that represents visual saliency of a corresponding visual scene. 4.
A saliency map model explains the effects of random variations along irrelevant dimensions in texture segmentation and visual search Li Zhaoping, University.
Attention, Awareness, and the Computational Theory of Surprise Research Qualifying Exam August 30 th, 2006.
Reading population codes: a neural implementation of ideal observers Sophie Deneve, Peter Latham, and Alexandre Pouget.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
BEYOND SIMPLE FEATURES: A LARGE-SCALE FEATURE SEARCH APPROACH TO UNCONSTRAINED FACE RECOGNITION Nicolas Pinto Massachusetts Institute of Technology David.
Michigan State University1 Visual Attention and Recognition Through Neuromorphic Modeling of “Where” and “What” Pathways Zhengping Ji Embodied Intelligence.
Fast, Multiscale Image Segmentation: From Pixels to Semantics Ronen Basri The Weizmann Institute of Science Joint work with Achi Brandt, Meirav Galun,
Eye Tracking for Personalized Photography Steven Scher (UCSC student) James Davis (UCSC advisor) Sriram Swaminarayan (LANL advisor)
Visual Attention More information in visual field than we can process at a given moment Solutions Shifts of Visual Attention related to eye movements Some.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 12: Visual Attention 1 Computational Architectures in Biological Vision,
1 Segmentation of Salient Regions in Outdoor Scenes Using Imagery and 3-D Data Gunhee Kim Daniel Huber Martial Hebert Carnegie Mellon University Robotics.
Visual Attention Jeremy Wyatt. Where to look? Many visual processes are expensive Humans don’t process the whole visual field How do we decide what to.
Christian Siagian Laurent Itti Univ. Southern California, CA, USA
Attention in Computer Vision Mica Arie-Nachimson and Michal Kiwkowitz May 22, 2005 Advanced Topics in Computer Vision Weizmann Institute of Science.
Overview 1.The Structure of the Visual Cortex 2.Using Selective Tuning to Model Visual Attention 3.The Motion Hierarchy Model 4.Simulation Results 5.Conclusions.
Extraction of Salient Contours in Color Images Vonikakis Vasilios, Ioannis Andreadis and Antonios Gasteratos Democritus University of Thrace
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
A neural mechanism for robust junction representation in the visual cortex University of Ulm Dept. of Neural Information Processing Thorsten Hansen and.
COLOR PERCEPTION Physical and Psychological Properties Theories – Trichromatic Theory – Opponent Process Theory Color Deficiencies Color and Lightness.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 5: Introduction to Vision 2 1 Computational Architectures in.
Rules by which the brain segments an object from the background: Evaluation of the Gabor model of simple cell receptive fields Beth Tuck Hanover College.
Michael Arbib & Laurent Itti: CS664 – Spring Lecture 5: Visual Attention (bottom-up) 1 CS 664, USC Spring 2002 Lecture 5. Visual Attention (bottom-up)
BATTLECAM™: A Dynamic Camera System for Real-Time Strategy Games Yangli Hector Yee Graphics Programmer, Petroglyph Elie Arabian.
Introduction of Saliency Map
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
CSNDSP’06 Visual Attention based Region of Interest Coding for Video - telephony Applications Nicolas Tsapatsoulis Computer Science Dept. University of.
Title of On the Implementation of a Information Hiding Design based on Saliency Map A.Basu, T. S. Das and S. K. Sarkar/ Jadavpur University/ Kolkata/ India/
Purdue University Page 1 Color Image Fidelity Assessor Color Image Fidelity Assessor * Wencheng Wu (Xerox Corporation) Zygmunt Pizlo (Purdue University)
Visual Attention Derek Hoiem March 14, 2007 Misc Reading Group.
黃文中 Introduction The Model Results Conclusion 2.
Assessment of Computational Visual Attention Models on Medical Images Varun Jampani 1, Ujjwal 1, Jayanthi Sivaswamy 1 and Vivek Vaidya 2 1 CVIT, IIIT Hyderabad,
Representations for object class recognition David Lowe Department of Computer Science University of British Columbia Vancouver, Canada Sept. 21, 2006.
The primate visual systemHelmuth Radrich, The primate visual system 1.Structure of the eye 2.Neural responses to light 3.Brightness perception.
Geodesic Saliency Using Background Priors
Department of Psychology & The Human Computer Interaction Program Vision Sciences Society’s Annual Meeting, Sarasota, FL May 13, 2007 Jeremiah D. Still,
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
Street Smarts: Visual Attention on the Go Alexander Patrikalakis May 13, XXX.
Efficient Color Boundary Detection with Color-opponent Mechanisms CVPR2013 Posters.
Binding problems and feature integration theory. Feature detectors Neurons that fire to specific features of a stimulus Pathway away from retina shows.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 12: Visual Attention 1 Computational Architectures in Biological.
Neural Models of Visual Attention John K. Tsotsos Center for Vision Research York University, Toronto, Canada Marc Pomplun Department of Computer Science.
A bottom up visual saliency map in the primary visual cortex --- theory and its experimental tests. Li Zhaoping University College London Invited presentation.
Spatio-temporal saliency model to predict eye movements in video free viewing Gipsa-lab, Grenoble Département Images et Signal CNRS, UMR 5216 S. Marat,
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
Cortical Representations of Attention and Salience James A. Mazer Department of Neurobiology Yale School of Medicine Theoretical and Experimental Approaches.
Sensation and Perception. Transformation of stimulus energy into a meaningful understanding –Each sense converts energy into awareness.
Biological Modeling of Neural Networks: Week 10 – Neuronal Populations Wulfram Gerstner EPFL, Lausanne, Switzerland 10.1 Cortical Populations - columns.
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
 Mentor : Prof. Amitabha Mukerjee Learning to Detect Salient Objects Team Members - Avinash Koyya Diwakar Chauhan.
Biologically Inspired Vision-based Indoor Localization Zhihao Li, Ming Yang
Activity Recognition Journal Club “Neural Mechanisms for the Recognition of Biological Movements” Martin Giese, Tomaso Poggio (Nature Neuroscience Review,
A Neurodynamical Cortical Model of Visual Attention and Invariant Object Recognition Gustavo Deco Edmund T. Rolls Vision Research, 2004.
Symbolic Reasoning in Spiking Neurons: A Model of the Cortex/Basal Ganglia/Thalamus Loop Terrence C. Stewart Xuan Choo Chris Eliasmith Centre for Theoretical.
Computational Vision --- a window to our brain
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
BATTLECAM™: A Dynamic Camera System for Real-Time Strategy Games
Implementation of a Visual Attention Model
Dynamic Causal Modelling (DCM): Theory
SIFT keypoint detection
by Khaled Nasr, Pooja Viswanathan, and Andreas Nieder
The functional architecture of attention
Presentation transcript:

Michigan State University 1 “Saliency-Based Visual Attention” “Computational Modeling of Visual Attention”, Itti, Koch, (Nature Reviews – Neuroscience 2001 ) “A Model of Saliency-Based Visual Attention for Rapid Scene Analysis”, Itti, Koch and Niebur’s (IEEE PAMI 1998) Zhengping Ji

Michigan State University 2 Overview l Background l System architecture l The saliency map l Preprocessing l Feature maps l Feature integration l Focus of attention l Results l Conclusion

Michigan State University 3 Related Work l “Feature Integration Theory,” Treisman & Gelade, l Computational model of bottom-up attention, Koch and Ullman, 1985 l Saliency map is believed to be located in the posterior parietal cortex (Gotlieb, et al., 1998) and the pulvinar nuclei of the thalamus (Roinson & Peterson, 1992)

Michigan State University 4 Architecture

Michigan State University 5 Gaussian Pyramids l Repeated low-pass filtering l =0, 1, 2, 3,…,8 I(0) is original input 640 x x x x60 Scaling by a factor 2x2 * G 5x5 Scaling by a factor 2x2 * G 5x5 *

Michigan State University 6 Preprocessing l Original image with red, green, blue channels Intensity as I = (r + g + b)/3 l Broadly tuned color channels R = r - (g + b)/2 G = g - (r + b)/2 B = b - (r + g)/2 Y = (r + g)/2 - |r – g|/2 - b

Michigan State University 7 Preprocessing RG B Y Intensity

Michigan State University 8 Center-surround Difference l Achieve center-surround difference through across-scale difference Operated denoted by  Interpolation to finer scale and point-to- point subtraction One pyramid for each channel: I(  ), R(  ), G(  ), B(  ), Y(  ) where   [0..8] is the scale

Michigan State University 9 Intensity Feature Maps I(c, s) = | I(c)  I(s) | c  {2, 3, 4} s = c +  where   {3, 4} So I (2, 5) = | I (2)  I (5)| I (2, 6) = | I (2)  I (6)| I (3, 6) = | I (3)  I (6)| … l  6 Feature Maps

Michigan State University 10 Colour Feature Maps Similar to double-opponent cells (Prim. V. C) Red-Green and Yellow-Blue RG(c, s) = | (R(c) - G(c))  (G(s) - R(s)) | BY(c, s) = | (B(c) - Y(c))  (Y(s) - B(s)) | Same c and s as with intensity +R-G +G-R +B-Y +Y-B

Michigan State University 11 Orientation Feature Maps Create Gabor pyramids for  = {0º, 45º, 90º, 135º} c and s again similar to intensity

Michigan State University 12 Normalization Operator l Promotes maps with few strong peaks l Surpresses maps with many comparable peaks 1. Normalization of map to range [ 0…M ] 2. Compute average m of all local maxima 3. Find the global maximum M 4. Multiply the map by ( M – m ) 2

Michigan State University 13 Normalization Operator

Michigan State University 14 Conspicuity Maps

Michigan State University 15 Saliency Map l Average all conspicuity maps

Michigan State University 16 Neural Layers l Saliency Map (SM) modeled as layer of leaky integrate-and-fire neurons l SM feeds into winner-take-all (WTA) neural network l Inhibition of Return as transient inhibition of SM at FOA SM Stimulus WTA Inhibition of Return FOA shifted to position of winner

Michigan State University 17 Example of Operation Inhibition of return

Michigan State University 18 Results Image Saliency Map High saliency Locations (yellow circles)

Michigan State University 19 Shifting Attention l Using 2D “winner- take-all” neural network at scale 4 l FOA shifts every ms

Michigan State University 20 Summary l Saliecy map can be broken down into main steps l Create pyramids for 5 channels of original image l Determine feature maps then conspicuity maps l Combine into saliency map (after normalizing) l The key idea of saliency map is to extract local spatial discontinuities in the modalities of color, intensity and orientation. l Use two layers of neurons to model shifting attention. l Model appears to work accurately and robustly (but difficult to evaluate)

Michigan State University 21 Discussion l No top-down attention modeling, e.g., top-down spacial control, obejct-based attention. l Biologically plausible? l Neuromorphic architecture? l In which way the top-down and bottom-up processes are related? l In which way the attention and recognition are integrated and interacted with each other?

Michigan State University 22 References Itti, Koch, and Niebur: “ A Model of Saliency-Based Visual Attention for Rapid Scene Analysis ” IEEE PAMI Vol. 20, No. 11, November (1998) Itti, Koch: “ Computational Modeling of Visual Attention ” Nature Reviews – Neuroscience Vol. 2 (2001)