Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.

Slides:



Advertisements
Similar presentations
Some Reflections on Augmented Cognition Eric Horvitz ISAT & Microsoft Research November 2000 Some Reflections on Augmented Cognition Eric Horvitz ISAT.
Advertisements

Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Introduction to Eye Tracking
Chapter 2: Marr’s theory of vision. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Introduce Marr’s distinction between.
Why do we move our eyes? - Image stabilization
Chapter Thirteen Conclusion: Where We Go From Here.
Perception Chapter 9: Event Perception Event Perception: an event is defined as a change in both time and space. Thus far we have discussed how our visual.
Detail to attention: Exploiting Visual Tasks for Selective Rendering Kirsten Cater 1, Alan Chalmers 1 and Greg Ward 2 1 University of Bristol, UK 2 Anyhere.
Perception and Perspective in Robotics Paul Fitzpatrick MIT Computer Science and Artificial Intelligence Laboratory Humanoid Robotics Group Goal To build.
Attention I Attention Wolfe et al Ch 7. Dana said that most vision is agenda-driven. He introduced the slide where the people attended to the many weird.
Control of Attention and Gaze in the Natural World.
Jeff B. Pelz Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of Technology Using Eyetracking to Improve Image Composition.
Correlation Between Image Reproduction Preferences and Viewing Patterns Measured with a Head Mounted Eye Tracker Lisa A. Markel Jeff B. Pelz, Ph.D. Center.
Disorders of Orienting Lesions to parietal cortex can produce some strange behavioural consequences –patients fail to notice events on the contralesional.
1 3 Processes of Pattern Recognition Sensation – you have to detect or see the pattern Perception – you have to organize the features into a whole Memory.
December 1, 2009Introduction to Cognitive Science Lecture 22: Neural Models of Mental Processes 1 Some YouTube movies: The Neocognitron Part I:
Saccades: Rapid rotation of the eyes to reorient gaze to a new object or region. Reflex saccades Attentional saccades Shifting Gaze - Saccades.
Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of.
Learning Sensorimotor Contingencies James J. Clark Centre for Intelligent Machines McGill University.
Summer 2011 Wednesday, 8/3. Biological Approaches to Understanding the Mind Connectionism is not the only approach to understanding the mind that draws.
Task Dependency of Eye Fixations & The Development of a Portable Eyetracker Jeff Cunningham Senior Research Project Dr. Jeff Pelz Visual Perception Laboratory.
Research Methods for HCI: Cognitive Modelling BCS HCI Tutorial 1 st September, 2008.
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
Eye Movements and Visual Attention
HUMAN 1 HUMAN 2 HUMAN 3 ALGORITHM ERRORS HUMAN 1 HUMAN 2 HUMAN 3 ALGORITHM ERRORS HUMAN 1 HUMAN 2 HUMAN 3 ALGORITHM ERRORS HUMAN 1 HUMAN 2 HUMAN 3 ALGORITHM.
The Eye-Tracking Butterfly: Morphing the SMI REDpt Eye-Tracking Camera into an Interactive Device. James Cunningham & James D. Miles California State University,
C OMPONENTS OF ISAVE Introduction Preparation Ecological Observations Structural Integrity Minimal Responsiveness Alignment and Ocular Mobility.
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
MIND: The Cognitive Side of Mind and Brain  “… the mind is not the brain, but what the brain does…” (Pinker, 1997)
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Visual Perception Is a Creative Process Instructor : Dr. S. Gharibzadeh Presented By : J. Razjouyan.
1 Perception and VR MONT 104S, Fall 2008 Session 13 Visual Attention.
Towards Cognitive Robotics Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Christian.
Eye movements: a primer Leanne Chukoskie, Ph.D.. Two classes of eye movements Gaze-stabilizing – Vestibulo-ocular reflex (VOR) – Optokinetic Nystagmus.
1 Computational Vision CSCI 363, Fall 2012 Lecture 20 Stereo, Motion.
Visual Perception, Attention & Action. Anthony J Greene2.
Natural Tasking of Robots Based on Human Interaction Cues Brian Scassellati, Bryan Adams, Aaron Edsinger, Matthew Marjanovic MIT Artificial Intelligence.
1 Computational Vision CSCI 363, Fall 2012 Lecture 36 Attention and Change Blindness (why you shouldn't text while driving)
Visually guided attention during flying OR Pilots “do not like” fovea because they cannot pay attention to more than 1% of space at any one time.
Visual Thinking and the Objects of Visual Attention Colin Ware University of New Hampshire.
Cognitive ability: Challenge: How to recognize objects in a scene; where are the object’s boundaries? This problem is known as ‘image segmentation’ in.
The geometry of the system consisting of the hyperbolic mirror and the CCD camera is shown to the right. The points on the mirror surface can be expressed.
Figure 20.1 Eye movements of a subject viewing a photograph of the bust of Queen Nefertiti neuro4e-fig jpg.
Research Background: Depth Exam Presentation
Mind, Brain & Behavior Wednesday February 19, 2003.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Computer Science Readings: Reinforcement Learning Presentation by: Arif OZGELEN.
1 Perception and VR MONT 104S, Fall 2008 Lecture 6 Seeing Motion.
Give examples of the way that virtual reality can be used in Psychology.
Development of Expertise. Expertise We are very good (perhaps even expert) at many things: - driving - reading - writing - talking What are some other.
Visual Perception By Katie Young and Joe Avery. Overview Visual Perception Eye Trackers Change Blindness.
How is vision used to catch a ball?
Copyright © 2007 Wolters Kluwer Health | Lippincott Williams & Wilkins Neuroscience: Exploring the Brain, 3e
Fundamentals of Sensation and Perception
Describe how reaching and grasping abilities develop in the first year of life.
Eye Movements and Working Memory Marc Pomplun Department of Computer Science University of Massachusetts at Boston Homepage:
What visual image information is needed for the things we do? How is vision used to acquire information from the world?
Scientific Advisory Group Meeting May 7, Scientific Advisory Group Meeting May 7, 2004 Access to Technical Education Through Sign Language Interpreting:
Functionality of objects through observation and Interaction Ruzena Bajcsy based on Luca Bogoni’s Ph.D thesis April 2016.
Chapter 11: Artificial Intelligence
Computer Vision Lecture 2: Vision, Attention, and Eye Movements
Movements of the eyes The muscles were of necessitie provided and given to the eye, that so it might move on every side; for if the eye stoode faste and.
Oculomotor Systems S. J. Potashner, Ph.D
Consequences of the Oculomotor Cycle for the Dynamics of Perception
Chapter 7 - Visual Attention
Consequences of the Oculomotor Cycle for the Dynamics of Perception
Eye-Based Interaction in Graphical Systems: Theory & Practice
Learning Sensorimotor Contingencies
Attention and Scene Perception
Presentation transcript:

Jeff B. Pelz, Roxanne Canosa, Jason Babcock, & Eric Knappenberger Visual Perception Laboratory Carlson Center for Imaging Science Rochester Institute of Technology Visual Perception in Familiar, Complex Tasks

© 2001 Rochester Institute of TechnologyVisual Perception RIT Students & other collaborators Students: Roxanne Canosa (Ph.D. Imaging Science) Jason Babcock (MS Color Science) Eric Knappenberger (MS Imaging Science) Collaborators: Mary Hayhoe UR Cognitive Science Dana Ballard UR Computer Science

© 2001 Rochester Institute of TechnologyVisual Perception RIT Visual Perception in Familiar, Complex Tasks Goals : è To b etter understand: Visual Perception and Cognition è To inform design of: AI and Computer Vision Systems

© 2001 Rochester Institute of TechnologyVisual Perception RIT A better understanding of the strategies and attentional mechanisms underlying visual perception and cognition in human observers can inform us on valuable approaches to computer-based perception. Strategic Vision

© 2001 Rochester Institute of TechnologyVisual Perception RIT å Sensorial Experience å High-level Visual Perception å Attentional Mechanisms å Eye Movements Motivation: Cognitive Science Visual Perception and Cognition The mechanisms underlying visual perception in humans are common to those needed to implement successful artificial vision systems.

© 2001 Rochester Institute of TechnologyVisual Perception RIT å Computer Vision å Active Vision å Attentional Mechanisms å Eye Movements Motivation: Computer Science Artificial Intelligence The mechanisms underlying visual perception in humans are common to those needed to implement successful artificial vision systems.

© 2001 Rochester Institute of TechnologyVisual Perception RIT As computing power and system sophistication grow, basic computer vision in constrained environments has become more tractable. But the value of computer vision systems will be shown in their ability to perform higher level actions in complex settings. Motivation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Even in the face of Moore’s Law, computers will not have sufficient power in the foreseeable future to solve the “vision problem” of image understanding by brute force. Limited Computational Resources

© 2001 Rochester Institute of TechnologyVisual Perception RIT Computer-based perception faces the same fundamental challenge that human perception did during evolution: limited computational resources Challenges

© 2001 Rochester Institute of TechnologyVisual Perception RIT Inspiration - “Active Vision” Active vision was the first step. Unlike traditional approaches to computer vision, active vision systems focused on extracting information from the scene rather than brute-force processing of static, 2D images.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Inspiration - “Active Vision” Visual routines were an important component of the approach. These pre- defined routines are scheduled and run to extract information when and where it is needed.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Goal - “Strategic Vision” “Strategic Vision” will focus on high-level, top-down strategies for extracting information from complex environments.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Goal - “Strategic Vision” A goal of our research is to study human behavior in natural, complex tasks because we are unlikely to identify these strategies in typical laboratory tasks that are usually used to study vision.

© 2001 Rochester Institute of TechnologyVisual Perception RIT In humans, “computational resources” means limited neural resources: even if the entire cortex were devoted to vision, there are not enough neurons to process and represent the full visual field at high acuity. Limited Computational Resources

© 2001 Rochester Institute of TechnologyVisual Perception RIT The solutions favored by evolution confronted the problem in three ways; 1. Anisotropic sampling of the scene 2. Serial execution 3. Limited internal representations Limited Neural Resources

© 2001 Rochester Institute of TechnologyVisual Perception RIT Retinal design: The “foveal compromise” design employs very high spatial acuity in a small central region (the fovea) coupled with a large field-of-view surround with limited acuity. periphery center periphery photoreceptor density 1. Anisotropic sampling of the scene

© 2001 Rochester Institute of TechnologyVisual Perception RIT Demonstration of the “ Foveal Compromise” Stare at the “+” below. Without moving your eyes, read the text presented in the next slide: +

© 2001 Rochester Institute of TechnologyVisual Perception RIT If you can read this you must be cheating + Anisotropic Sampling: Foveal Compromise

© 2001 Rochester Institute of TechnologyVisual Perception RIT Anisotropic Sampling: Foveal Compromise Despite the conscious percept of a large, high-acuity field-of-view, only a small fraction of the field is represented with sufficient fidelity for tasks requiring even moderate acuity.

© 2001 Rochester Institute of TechnologyVisual Perception RIT The solutions favored by evolution confronted the problem in three ways; 1. Anisotropic sampling of the scene 2. Serial execution 3. Limited internal representations Limited Neural Resources

© 2001 Rochester Institute of TechnologyVisual Perception RIT The limited acuity periphery must be sampled by the high-acuity fovea. This sampling imposes a serial flow of information, with successive views. Serial Execution

© 2001 Rochester Institute of TechnologyVisual Perception RIT Extraocular Muscles: Three pairs of extra-ocular muscles provide a rich suite of eye movements that can rapidly move the point of gaze to sample the scene or environment with the high- acuity central fovea. Serial Execution: Eye Movements

© 2001 Rochester Institute of TechnologyVisual Perception RIT Background: Eye Movement Types Image stabilization during object and/or observer motion Smooth pursuit/ Optokinetic response Vestibular-ocular response Vergence Saccades Image destabilization - used to shift gaze to new locations.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Background: Eye Movement Types Saccadic Eye Movements: Rapid, ballistic eye movements that move the eye from point to point in the scene Separated by fixations during which the retinal image is stabilized

© 2001 Rochester Institute of TechnologyVisual Perception RIT Serial Execution; Image Preference 3 sec viewing

© 2001 Rochester Institute of TechnologyVisual Perception RIT Serial Execution: Complex Tasks Saccadic eye movements are also seen in complex tasks

© 2001 Rochester Institute of TechnologyVisual Perception RIT The solutions favored by evolution confronted the problem in three ways; 1. Anisotropic sampling of the scene 2. Serial execution 3. Limited internal representations Limited Neural Resources

© 2001 Rochester Institute of TechnologyVisual Perception RIT How are successive views integrated? What fidelity is the integrated representation? Integration of Successive Fixations

© 2001 Rochester Institute of TechnologyVisual Perception RIT [fixate the cross] Integration of Successive Fixations

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Integration of Successive Fixations Perhaps fixations are integrated in an internal representation

© 2001 Rochester Institute of TechnologyVisual Perception RIT Limited Representation: “Change blindness” B A If successive fixations are used to build up a high-fidelity internal representation, then it should be easy to detect even small differences between two images.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image A Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image B Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image A Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image B Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image A Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Image B Try to identify the difference between Image A & B

© 2001 Rochester Institute of TechnologyVisual Perception RIT Can we identify oculomotor strategies that observers use to ease the computational and memory load on observers as they perceive the real world? The question:

© 2001 Rochester Institute of TechnologyVisual Perception RIT To answer that question, we have to monitor eye movements in the real world, as people perform real extended tasks. One problem is the hardware: The question:

© 2001 Rochester Institute of TechnologyVisual Perception RIT Measuring eye movements Scleral eye-coils Dual Purkinje eyetracker Many eyetracking systems require that head move- ments (and other natural behaviors) be restricted.

© 2001 Rochester Institute of TechnologyVisual Perception RIT RIT Wearable Eyetracker The RIT wearable eyetracker is a self-contained unit that allows monitoring subjects’ eye movements during natural tasks.

© 2001 Rochester Institute of TechnologyVisual Perception RIT RIT Wearable Eyetracker The headgear holds CMOS cameras and IR source. Controller and video recording devices are held in a backpack.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Beyond the mechanics of how the eyes move during real tasks, we are interested in strategies that may support the conscious perception that is continuous temporally as well as spatially. Perceptual strategies

© 2001 Rochester Institute of TechnologyVisual Perception RIT When subjects’ eye movements are monitored as they perform familiar complex tasks, novel sequences of eye movements were seen that demonstrate strategies that simplify the perceptual load of these tasks. Perceptual strategies

© 2001 Rochester Institute of TechnologyVisual Perception RIT In laboratory tasks, subjects usually fixate objects of immediate relevance. When we monitor subjects performing complex tasks, we also observe fixations on objects that were relevant only to future interactions. The next slide shows a series of fixations as a subject approached a sink and washed his hands: Perceptual strategies

© 2001 Rochester Institute of TechnologyVisual Perception RIT Early fixations are on the water faucet; the first object for interaction. t = -700 msec t = 0 msec Even before the faucet is reached, the subject looks ahead to the soap dispenser. “Look-ahead” Fixations t = 1500 msec Fixations return to items of immediate interaction. t = 2000 msec Gaze returns again to the dispenser just before the reach to it begins. t = 2600 msec Gaze remains on the dispenser until the hand reaches it. t = 2800 msec

© 2001 Rochester Institute of TechnologyVisual Perception RIT guiding fixation look-ahead fixationinteraction 2000 msec800 msec Perceptual Strategies: Where, when, why? The “look-ahead” fixations are made in advance of “guiding” fixations used before interacting with objects in the environment.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Conclusion Humans employ strategies to ease the computational and memory loads inherent in complex tasks. Look-ahead fixations reveal one such strategy: opportunistic execution of information- gathering routines to pre-fetch information needed for future subtasks.

© 2001 Rochester Institute of TechnologyVisual Perception RIT Future Work Future work will implement this form of opportunistic execution in artificial vision systems to test the hypothesis that strategic visual routines observed in humans can benefit computer-based perceptual systems.