A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John.

Slides:



Advertisements
Similar presentations
Read this article for Friday next week [1]Chelazzi L, Miller EK, Duncan J, Desimone R. A neural basis for visual search in inferior temporal cortex. Nature.
Advertisements

Central Visual Processes. Anthony J Greene2 Central Visual Pathways I.Primary Visual Cortex Receptive Field Columns Hypercolumns II.Spatial Frequency.
Object recognition and scene “understanding”
September 2, 2014Computer Vision Lecture 1: Human Vision 1 Welcome to CS 675 – Computer Vision Fall 2014 Instructor: Marc Pomplun Instructor: Marc Pomplun.
Human (ERP and imaging) and monkey (cell recording) data together 1. Modality specific extrastriate cortex is modulated by attention (V4, IT, MT). 2. V1.
MOTION PERCEPTION Types of Motion Perception Corollary Discharge Theory Movement Detectors Motion Perception and Object Perception Ecological Perception.
Visual Neuron Responses This conceptualization of the visual system was “static” - it did not take into account the possibility that visual cells might.
1 Biological Neural Networks Example: The Visual System.
How does the visual system represent visual information? How does the visual system represent features of scenes? Vision is analytical - the system breaks.
[1].Edward G. Jones, Microcolumns in the Cerebral Cortex, Proc. of National Academy of Science of United States of America, vol. 97(10), 2000, pp
Introduction to Cognitive Science Lecture 2: Vision in Humans and Machines 1 Vision in Humans and Machines September 10, 2009.
COGNITIVE NEUROSCIENCE
EE141 1 Broca’s area Pars opercularis Motor cortexSomatosensory cortex Sensory associative cortex Primary Auditory cortex Wernicke’s area Visual associative.
Brain Theory and Artificial Intelligence
December 1, 2009Introduction to Cognitive Science Lecture 22: Neural Models of Mental Processes 1 Some YouTube movies: The Neocognitron Part I:
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 12: Visual Attention 1 Computational Architectures in Biological Vision,
VISUAL PATHWAYS Organization of LGN of thalamus Organization of Visual Cortex What Stream How Stream The Binding Problem.
Michigan State University 1 “Saliency-Based Visual Attention” “Computational Modeling of Visual Attention”, Itti, Koch, (Nature Reviews – Neuroscience.
The Human Visual System Vonikakis Vasilios, Antonios Gasteratos Democritus University of Thrace
IT 691 Final Presentation Pace University Created by: Robert M Gust Mark Lee Samir Hessami Mark Lee Samir Hessami.
Overview 1.The Structure of the Visual Cortex 2.Using Selective Tuning to Model Visual Attention 3.The Motion Hierarchy Model 4.Simulation Results 5.Conclusions.
Motion Computation and Visual Orientation In Flies MSc Evolutionary and Adaptive Systems Computer Vision Juan Pablo Calderon.
Chapter 10 The Central Visual System. Introduction Neurons in the visual system –Neural processing results in perception Parallel pathway serving conscious.
October 7, 2010Neural Networks Lecture 10: Setting Backpropagation Parameters 1 Creating Data Representations On the other hand, sets of orthogonal vectors.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 5: Introduction to Vision 2 1 Computational Architectures in.
Michael Arbib & Laurent Itti: CS664 – Spring Lecture 5: Visual Attention (bottom-up) 1 CS 664, USC Spring 2002 Lecture 5. Visual Attention (bottom-up)
University Studies 15A: Consciousness I The Neurobiology of Vision.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Text Lecture 2 Schwartz.  The problem for the designer is to ensure all visual queries can be effectively and rapidly served.  Semantically meaningful.
Low Level Visual Processing. Information Maximization in the Retina Hypothesis: ganglion cells try to transmit as much information as possible about the.
Advances in Modeling Neocortex and its impact on machine intelligence Jeff Hawkins Numenta Inc. VS265 Neural Computation December 2, 2010 Documentation.
1 Computational Vision CSCI 363, Fall 2012 Lecture 31 Heading Models.
THE VISUAL SYSTEM: EYE TO CORTEX Outline 1. The Eyes a. Structure b. Accommodation c. Binocular Disparity 2. The Retina a. Structure b. Completion c. Cone.
Slide 1 Neuroscience: Exploring the Brain, 3rd Ed, Bear, Connors, and Paradiso Copyright © 2007 Lippincott Williams & Wilkins Bear: Neuroscience: Exploring.
黃文中 Introduction The Model Results Conclusion 2.
1 Perception, Illusion and VR HNRS 299, Spring 2008 Lecture 8 Seeing Depth.
The primate visual systemHelmuth Radrich, The primate visual system 1.Structure of the eye 2.Neural responses to light 3.Brightness perception.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Biological Modeling of Neural Networks: Week 12 – Decision models: Competitive dynamics Wulfram Gerstner EPFL, Lausanne, Switzerland 12.1 Review: Population.
Human vision Jitendra Malik U.C. Berkeley. Visual Areas.
Lecture 21 Neural Modeling II Martin Giese. Aim of this Class Account for experimentally observed effects in motion perception with the simple neuronal.
September 3, 2013Computer Vision Lecture 1: Human Vision 1 Welcome to CS 675 – Computer Vision Fall 2013 Instructor: Marc Pomplun Instructor: Marc Pomplun.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 12: Visual Attention 1 Computational Architectures in Biological.
1 Perception and VR MONT 104S, Fall 2008 Lecture 6 Seeing Motion.
Colour and Texture. Extract 3-D information Using Vision Extract 3-D information for performing certain tasks such as manipulation, navigation, and recognition.
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Week 4 Motion, Depth, Form: Cormack Wolfe Ch 6, 8 Kandell Ch 27, 28 Advanced readings: Werner and Chalupa Chs 49, 54, 57.
Sensation & Perception. Motion Vision I: Basic Motion Vision.
Cogs1 mapping space in the brain Douglas Nitz – Feb. 19, 2009 any point in space is defined relative to other points in space.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
1 Computational Vision CSCI 363, Fall 2012 Lecture 32 Biological Heading, Color.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
Activity Recognition Journal Club “Neural Mechanisms for the Recognition of Biological Movements” Martin Giese, Tomaso Poggio (Nature Neuroscience Review,
National Taiwan Normal A System to Detect Complex Motion of Nearby Vehicles on Freeways C. Y. Fang Department of Information.
A Neurodynamical Cortical Model of Visual Attention and Invariant Object Recognition Gustavo Deco Edmund T. Rolls Vision Research, 2004.
Implementation of a Visual Attention Model
Computer Vision Lecture 2: Vision, Attention, and Eye Movements
Welcome to CS 675 – Computer Vision Spring 2018
CORTICAL MECHANISMS OF VISION
Creating Data Representations
Consequences of the Oculomotor Cycle for the Dynamics of Perception
Motion vision: Are ‘speed lines’ used in human visual motion?
Consequences of the Oculomotor Cycle for the Dynamics of Perception
The Normalization Model of Attention
Learning Sensorimotor Contingencies
Detecting Motion Pattern in Optical Flow Fields
Presentation transcript:

A Neural Model for Detecting and Labeling Motion Patterns in Image Sequences Marc Pomplun 1 Julio Martinez-Trujillo 2 Yueju Liu 2 Evgueni Simine 2 John Tsotsos 2 1 UMass Boston 2 York University, Toronto, Canada

“ Data Flow Diagram ” of Visual Areas in Macaque Brain Blue: motion perception pathway Green: object recognition pathway

Receptive Fields in Hierarchical Neural Networks neuron A receptive field of A

Receptive Fields in Hierarchical Neural Networks receptive field of A in input layer neuron A in top layer

poor localization crosstalk Problems with Information Routing in Hierarchical Networks

The Selective Tuning Concept (Tsotsos, 1988) processing pyramid inhibited pathways pass pathways: hierarchical restriction of input space unit of interest at top input

top-down, coarse-to-fine WTA hierarchy for selection and localization unselected connections are inhibited WTA achieved through local gating networks Hierarchical Winner-Take-All

unit and connection in the interpretive network unit and connection in the gating network unit and connection in the top-down bias network layer +1 layer  -1 layer I Selection Circuits

3D Visualization of the Selective Tuning Network Red: WTA phase 1 activeGreen: WTA phase 2 active Blue: inhibitionYellow: WTA winner

The Motion Perception Pathway MST MT V1 feed- forward feedback input feed- forward feedback

What do We Know about Area V1? cells have small receptive fields each cell has a preferred direction of motion direction of motion activation preferred direction there are three types of motion speed selectivity speed of motion activation low-speed cells medium-speed cells high-speed cells

What do We Know about Area MT? cells have larger receptive fields than in V1 like in V1, each cell has a preferred combination of the direction and speed of motion MT cells also have a preferred orientation of the speed gradient orientation of speed gradient activation preferred orientation of speed gradient without speed gradient with speed gradient

What do We Know about Area MST? cells respond to motion patterns such as –translation (objects shifting positions) –rotation (clockwise and counterclockwise) –expansion (approaching objects) –contraction (receding objects) –spiral motion (combinations of rotation and expansion/contraction) the response of a cell is almost independent on the position of the motion pattern in the visual field

The Motion Hierarchy Model: V1 V1 receives image sequences as input and extracts the direction and speed of motion counterclockwise rotationclockwise rotationcontractionexpansion counterclockwiseclockwise contractionexpansion

The Motion Hierarchy Model: V1 V1 is simulated as 60x60 hypercolumns each column contains 36 cells: one for each combination of direction (12) and speed tuning (3) direction and speed selectivity are achieved with spatiotemporal filters these filters process local information from the last seven images in the sequence example: cells tuned towards upward motion: input pattern: counter-clockwise rotation high-speed cells medium-speed cells low-speed cells

The Motion Hierarchy Model: MT MT is simulated as 30x30 hypercolumns each column contains 432 cells: one for each combination of direction (12) speed (3), and speed gradient tuning (12) problem: how can gradient tuning be realized from activation patterns in V1? –solution: detect gradient differences across the three types of speed selective cells –this solution leads to a simple network structure and remarkably good noise reduction the activation of an MT cell is the product of its activation by direction, speed, and gradient

The Motion Hierarchy Model: MST how can MST cells detect motion patterns such as rotation, expansion, and contraction based on the activation of MT cells? counterclockwiseclockwisecontractionexpansion movementspeed gradient idea: the presence of these motion patterns is indicated by a consistent angle between the local movement and speed gradient

The Motion Hierarchy Model: MST direction of movement orientation of speed gradient

The Motion Hierarchy Model: MST MST cells integrate the activation of MT cells that respond to a particular angle between motion and speed gradient this integration is performed across a large part of the visual field and across all 12 directions therefore, MST can detect 12 different motion patterns we simulate 5x5 MST hypercolumns, each containing 36 neurons (tuned for 12 different motion patterns, 3 different speeds)

direction of movement speed gradient V1 MT MST

Simulation: clockwise rotation direction of movement speed gradient

Simulation: counter- clockwise rotation direction of movement speed gradient

Simulation: receding object direction of movement speed gradient

Attention in the Motion Hierarchy What happens if there are multiple motion patterns in the visual input? Visual attention can be used to determine the type and location of the most salient motion pattern, focus on it by eliminating all interfering information, sequentially inspect all objects in the visual field.

direction of movement speed gradient

direction of movement speed gradient

Conclusions and Outlook the motion hierarchy model provides a plausible explanation for cell properties in areas V1, MT, and MST its use of distinct speed tuning functions in V1 and speed gradient selectivity in MT leads to a relatively simple network structure combined with robust and precise detection of motion patterns visual attention is employed to segregate and sequentially inspect multiple motion patterns

Conclusions and Outlook the model predicts inhibition of visual functions around any attended motion pattern the model also predicts that different motion patterns induce different activation patterns in V1, MT, and MST linear motion activates V1, MT, and MST speed gradients increase MT and MST activation rotation, expansion, and contraction increase MST activation this is currently being tested by fMRI scanning experiments in Magdeburg, Germany

Conclusions and Outlook the model is well-suited for mobile robots to estimate parameters of ego-motion the area MST in the simulated hierarchy is very sensitive to any translational or rotational ego-motion in biological vision, MST is massively connected to the vestibular system in mobile robots, the simulated area MST could interact with position and orientation sensors to stabilize ego- motion estimation

Conclusions and Outlook Future work: lateral interaction across neighboring sets of gating units for improved perceptual grouping simultaneous simulation of both the motion perception and object recognition pathways introduction of working memory for an adequate internal representation of the current visual scene