Major Functional Areas Primary motor: voluntary movement Primary somatosensory: tactile, pain, pressure, position, temp., mvt. Motor association: coordination.

Slides:



Advertisements
Similar presentations
Chapter 4: The Visual Cortex and Beyond
Advertisements

Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Image Registration  Mapping of Evolution. Registration Goals Assume the correspondences are known Find such f() and g() such that the images are best.
Object recognition and scene “understanding”
Why do we move our eyes? - Image stabilization
Virtual Reality Design Virtual reality systems are designed to produce in the participant the cognitive effects of feeling immersed in the environment.
Human (ERP and imaging) and monkey (cell recording) data together 1. Modality specific extrastriate cortex is modulated by attention (V4, IT, MT). 2. V1.
SAL: A Hybrid Cognitive Architecture Y. Vinokurov 1, C. Lebiere 1, D. Wyatte 2, S. Herd 2, R. O’Reilly 2 1. Carnegie Mellon University, 2. University of.
2002/02/05PSYC , Term 2, Copyright Jason Harrison, Visual pathways from here to there to everywhere.
The Human Brain.
The Brain Notes.
The Brain! IIA-3.1 Identify the structure and function of the major regions of the brain. IIA-3.2 Recognize that specific functions are centered in specific.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Robust Object Tracking via Sparsity-based Collaborative Model
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Visual Search: finding a single item in a cluttered visual scene.
Read this article for Friday next week [1]Chelazzi L, Miller EK, Duncan J, Desimone R. A neural basis for visual search in inferior temporal cortex. Nature.
MENTAL REPRESENATIONS Neur 3680 Midterm I review.
Read Lamme (2000) TINS article for Wednesday. Visual Pathways V1 is, of course, not the only visual area (it turns out it’s not even always “primary”)
How does the visual system represent visual information? How does the visual system represent features of scenes? Vision is analytical - the system breaks.
Computational Vision: Object Recognition Object Recognition Jeremy Wyatt.
A Lightweight Computer- Vision-based Electronic Travel Aid Andrew B. Raij Enabling Tech Project Final Report 4/17/2003.
Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.
Tracking multiple independent targets: Evidence for a parallel tracking mechanism Zenon Pylyshyn and Ron Storm presented by Nick Howe.
CS 664, Session 19 1 General architecture. CS 664, Session 19 2 Minimal Subscene Working definition: The smallest set of objects, actors and actions in.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 12: Visual Attention 1 Computational Architectures in Biological Vision,
Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.
CS 664, Session 20 1 Minimal Subscene Working definition: The smallest set of objects, actors and actions in a dynamic visual scene that are relevant to.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 2: Neuroscience Basics 1 Computational Architectures in Biological.
Summer 2011 Wednesday, 8/3. Biological Approaches to Understanding the Mind Connectionism is not the only approach to understanding the mind that draws.
Laurent Itti: CS564 – Brain theory and artificial intelligence. Scene Perception 1 Brain theory and artificial intelligence Lecture 23. Scene Perception.
Jochen Triesch, UC San Diego, 1 COGS Visual Modeling Jochen Triesch & Martin Sereno Dept. of Cognitive Science UC.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 2: Neuroscience Basics 1 Computational Architectures in Biological.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC. Lecture 13: Scene Perception 1 Computational Architectures in Biological Vision,
Overview of upcoming lectures 1/15: General introduction to vision, brain and the “minimal subscene” 1/17: Overview and discussion of projects 1/22: Visual.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 13: Scene Perception 1 Computational Architectures in Biological.
Sana Naghipour, Saba Naghipour Mentor: Phani Chavali Advisers: Ed Richter, Prof. Arye Nehorai.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
MIND: The Cognitive Side of Mind and Brain  “… the mind is not the brain, but what the brain does…” (Pinker, 1997)
Computer graphics & visualization REYES Render Everything Your Eyes Ever Saw.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Biases: An Example Non-accidental properties: Properties that appear in an image that are very unlikely to have been produced by chance, and therefore.
Visual Perception Is a Creative Process Instructor : Dr. S. Gharibzadeh Presented By : J. Razjouyan.
1 Perception and VR MONT 104S, Fall 2008 Session 13 Visual Attention.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
The Visual Cortex: Anatomy
Visual Attention Derek Hoiem March 14, 2007 Misc Reading Group.
Spatio-temporal constraints for recognizing 3D objects in videos Nicoletta Noceti Università degli Studi di Genova.
Generalized Hough Transform
Nervous System: Part VI Specialized Receptors: Eyes and Ears.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Object Recognition in ROS Using Feature Extractors and Feature Matchers By Dolev Shapira.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Monitoring and Enhancing Visual Features (movement, color) as a Method for Predicting Brain Activity Level -in Terms of the Perception of Pain Sensation.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 12: Visual Attention 1 Computational Architectures in Biological.
Give examples of the way that virtual reality can be used in Psychology.
How conscious experience and working memory interact Bernard J. Baars and Stan Franklin Soft Computing Laboratory 김 희 택 TRENDS in Cognitive Sciences vol.
Optimal Eye Movement Strategies In Visual Search.
1 Copyright © 2014 Elsevier Inc. All rights reserved. Chapter 19 Visual Network Moran Furman.
Sensation and Perception Unit 5 Meyers: Ch. 5,6 Barrons: Ch. 4 Passer/Smith: Ch. 5.
Chapter 15. Cognitive Adequacy in Brain- Like Intelligence in Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Cinarel, Ceyda.
Activity Recognition Journal Club “Neural Mechanisms for the Recognition of Biological Movements” Martin Giese, Tomaso Poggio (Nature Neuroscience Review,
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching Link: singhashwini.mesinghashwini.me.
Assist. Prof. Dr. Ilmiye Seçer Fall
Recognizing Deformable Shapes
Perception Unit How can we so easily perceive the world? Objects may be moving, partially hidden, varying in orientation, and projected as a 2D image.
Course Instructor: knza ch
The Visual System: Higher Cortical Mechanisms
Presentation transcript:

Major Functional Areas Primary motor: voluntary movement Primary somatosensory: tactile, pain, pressure, position, temp., mvt. Motor association: coordination of complex movements Sensory association: processing of multisensorial information Prefrontal: planning, emotion, judgement Speech center (Broca’s area): speech production and articulation Wernicke’s area: comprehen- sion of speech Auditory: hearing Auditory association: complex auditory processing Visual: low-level vision Visual association: higher-level vision

“Where” and “What” Visual Pathways Dorsal stream (to posterior parietal): object localization Ventral stream (to infero-temporal): object identification Rybak et al, 1998 “where” “what” “where” “what”

Riesenhuber & Poggio, Nat Neurosci, 1999 The next step… Develop scene understanding/navigation/orienting mechanisms that can exploit the (very noisy) “rich scanpaths” (i.e., with location and sometimes identification) generated by the model.

Extract “minimal subscene” (i.e., small number of objects and actions) that is relevant to present behavior. Achieve representation for it that is robust and stable against noise, world motion, and egomotion.

How plausible is that? 3—5 eye movements/sec, that’s 150,000—250,000/day Only central 2deg of retinas (our foveas) carry high- resolution information Attention helps us bring our foveas onto relevant objects.

Eye Movements 1) Free examination 2) estimate material circumstances of family 3) give ages of the people 4) surmise what family has been doing before arrival of “unexpected visitor” 5) remember clothes worn by the people 6) remember position of people and objects 7) estimate how long the “unexpected visitor” has been away from family Yarbus, 1967

Extended Scene Perception Attention-based analysis: Scan scene with attention, accumulate evidence from detailed local analysis at each attended location. Main issues: - what is the internal representation? - how detailed is memory? - do we really have a detailed internal representation at all?

Rybak et al, 1998

Algorithm -At each fixation, extract central edge orientation, as well as a number of “context” edges; -Transform those low-level features into more invariant “second order” features, represented in a referential attached to the central edge; -Learning: manually select fixation points; store sequence of second-order features found at each fixation into “what” memory; also store vector for next fixation, based on context points and in the second-order referential;

Algorithm As a result, sequence of retinal images is stored in “what” memory, and corresponding sequence of attentional shifts in the “where” memory.

Algorithm -Search mode: look for an image patch that matches one of the patches stored in the “what” memory; -Recognition mode: reproduce scanpath stored in memory and determine whether we have a match.

Robust to variations in scale, rotation, illumination, but not 3D pose.

Schill et al, JEI, 2001

The World as an Outside Memory Kevin O’Regan, early 90s: why build a detailed internal representation of the world? too complex… not enough memory… … and useless? The world is the memory. Attention and the eyes are a look-up tool.

The “Attention Hypothesis” No “integrative buffer” Early processing extracts information up to “proto-object” complexity in massively parallel manner Attention is necessary to bind the different proto-objects into complete objects, as well as to bind object and location Once attention leaves an object, the binding “dissolves.” Not a problem, it can be formed again whenever needed, by shifting attention back to the object. Only a rather sketchy “virtual representation” is kept in memory, and attention/eye movements are used to gather details as needed Rensink, 2000

Outlook on human vision Unlikely that we perceive scenes by building a progressive buffer and accumulating detailed evidence into it. - too much resources - too complex to use. Rather, we may only have an illusion of detailed representation. - use eyes/attention to get the details as needed - the world as an outside memory. In addition to attention-based scene analysis, we are able to very rapidly extract the gist & layout of a scene – much faster than we can shift attention around. This gist/layout must be constructed by fairly simple processes that operate in parallel. It can then be used to prime memory and attention.

Goal-directed scene understanding Goal: develop vision/language-enabled AI system. Architecture it after the primate brain Test: ask a question to system about a video clip that it is watching e.g., “Who is doing what to whom?” Test: implement system on mobile robot and give it some instructions e.g., “Go to the library”

Applications include… -Autonomous vehicles / helper robots -Warning/helper systems on board of human-operated machinery and vehicles - Human/computer interfaces -Vision-based aids for the visually impaired

Example Question: “who is doing what to whom?” Answer: “Eric passes, turns around and passes again”

General architecture

Example Question: “who is doing what to whom?” Answer: “Eric passes, turns around and passes again”

Outlook Neuromorphic vision algorithms provide robust front-end for extended scene analysis To be useful, analysis needs to highly depend on task and behavioral priorities Thus the challenge is to develop algorithms that can extract the currently relevant “minimal subscene” from incoming rich scanpaths Such algorithms will use a collaboration between fast parallel computation of scene gist/layout and slower attentional scanning of scene details

“Beobot”

Eric Pichon Nathan Mundhenk Daesu Chung April Tsui Tasha Drew Jen Ng Ruchi Modi Yoohee Shin Reid Hirata Tony Ventrice