Control of Attention and Gaze in Natural Environments.

Slides:



Advertisements
Similar presentations
Introduction to Eye Tracking
Advertisements

All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Effect of Opacity of Stimulus in Deployment of Interest in an Interface Sujoy Kumar Chowdhury & Jeremiah D. Still Missouri Western State University Introduction.
Detail to attention: Exploiting Visual Tasks for Selective Rendering Kirsten Cater 1, Alan Chalmers 1 and Greg Ward 2 1 University of Bristol, UK 2 Anyhere.
Modeling the Brain’s Operating System Dana H. Ballard Computer Science Dept. University of Austin Texas, NY, USA International Symposium “Vision by Brains.
Chapter 6: Visual Attention. Overview of Questions Why do we pay attention to some parts of a scene but not to others? Do we have to pay attention to.
Eye Movements of Younger and Older Drivers Professor: Liu Student: Ruby.
LOGO The role of attentional breadth in perceptual change detection Professor: Liu Student: Ruby.
Psych 216: Movement Attention. What is attention? There is too much information available in the world to process it all. Demonstration: change-detection.
Spatial Neglect and Attention Networks
Attention I Attention Wolfe et al Ch 7. Dana said that most vision is agenda-driven. He introduced the slide where the people attended to the many weird.
Vision Based Control Motion Matt Baker Kevin VanDyke.
Chapter 6: Visual Attention. Scanning a Scene Visual scanning – looking from place to place –Fixation –Saccadic eye movement Overt attention involves.
1 Reactive Pedestrian Path Following from Examples Ronald A. Metoyer Jessica K. Hodgins Presented by Stephen Allen.
Control of Attention and Gaze in the Natural World.
Mobile Phone Use in a Driving Simulation Task: Differences in Eye Movements Stacy Balk, Kristin Moore, Will Spearman, & Jay Steele.
Current Trends in Image Quality Perception Mason Macklem Simon Fraser University
The Use of Eye Tracking Technology in the Evaluation of e-Learning: A Feasibility Study Dr Peter Eachus University of Salford.
Gray et al., The Soft Constraints Hypothesis: A Rational Analysis Approach to Resource Allocation for Interactive Behavior.
Brian White, Karl Gegenfurtner & Dirk Kerzel Random noise textures as visual distractors.
Baysian Approaches Kun Guo, PhD Reader in Cognitive Neuroscience School of Psychology University of Lincoln Quantitative Methods 2011.
December 2, 2014Computer Vision Lecture 21: Image Understanding 1 Today’s topic is.. Image Understanding.
EMBEDDED SOFTWARE Team victorious Team Victorious.
Manipulating Attention in Computer Games Matthias Bernhard, Le Zhang, Michael Wimmer Institute of Computer Graphics and Algorithms Vienna University of.
Cognitive demands of hands-free- phone conversation while driving Professor : Liu Student: Ruby.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
LOGO A review of individual differences in field dependence as a factor in auto safety Professor: Liu Student: Ruby.
Eye movements: Lab # 1 - Catching a ball. How do we use our eyes to catch balls? What information does the brain need? Most experiments look at simple.
1 Perception and VR MONT 104S, Fall 2008 Session 13 Visual Attention.
Logo Add Your Company Slogan A field evaluation of driver eye and head movement strategies toward environmental targets and distracters Professor: Liu.
Methods Inhibition of Return was used as a marker of attention capture.  After attention goes to a location it is inhibited from returning later. Results.
Describe 2 kinds of eye movements and their function. Describe the specialized gaze patterns found by Land in cricket. Describe your results in the ball-catching.
Subject wearing a VR helmet immersed in the virtual environment on the right, with obstacles and targets. Subjects follow the path, avoid the obstacels,
Change detection and occlusion modes in road-traffic scenarios Professor: Liu Student: Ruby.
Psych 435 Attention. Issues Capacity –We can’t respond to everything in the environment –Too many pieces of information –we can only actively respond.
1 Computational Vision CSCI 363, Fall 2012 Lecture 36 Attention and Change Blindness (why you shouldn't text while driving)
Seeing and Acting in a Virtual World PSY 341K Class hours: Tues, Thurs Room 4-242, SEAY Instructor: Professor Mary Hayhoe SEAY Room X
Experiment 2 (N=10) Purpose: Examine the ability of rare abrupt onsets (20% of trials) to capture attention away from a relevant cue. Design: Half of the.
The effects of working memory load on negative priming in an N-back task Ewald Neumann Brain-Inspired Cognitive Systems (BICS) July, 2010.
Summary of results. Reiterate goal of investigation: How general is anticipatory behavior observed by Land & McCleod? Found: Untrained subjects exhibit.
Object Lesson: Discovering and Learning to Recognize Objects Object Lesson: Discovering and Learning to Recognize Objects – Paul Fitzpatrick – MIT CSAIL.
The geometry of the system consisting of the hyperbolic mirror and the CCD camera is shown to the right. The points on the mirror surface can be expressed.
Computer Science Readings: Reinforcement Learning Presentation by: Arif OZGELEN.
Psych 335 Attention. Issues Capacity –We can’t respond to everything in the environment –Too many pieces of information –we can only actively respond.
Older Driver Failures of Attention at Intersections: Using Change Blindness Methods to Assess Turn Decision Accuracy Professor: Liu Student: Ruby.
An Eyetracking Analysis of the Effect of Prior Comparison on Analogical Mapping Catherine A. Clement, Eastern Kentucky University Carrie Harris, Tara Weatherholt,
How is vision used to catch a ball?
LOGO Change blindness in the absence of a visual disruption Professor: Liu Student: Ruby.
 Example: seeing a bird that is singing in a tree or miss a road sign in plain sight  Cell phone use while driving reduces attention and memory for.
8. What are the advantages and disadvantages of using a virtual reality environment to study the brain and behavior? 9.Give examples of the way that virtual.
Picture change during blinks: looking without seeing and seeing without looking Professor: Liu Student: Ruby.
Adaptive Control of Gaze and Attention Mary Hayhoe University of Texas at Austin Jelena Jovancevic University of Rochester Brian Sullivan University of.
What is meant by “top-down” and “bottom-up” processing? Give examples of both. Bottom up processes are evoked by the visual stimulus. Top down processes.
Control of Attention and Gaze in Natural Environments.
Describe how reaching and grasping abilities develop in the first year of life.
Traffic scene related change blindness in older drivers Professor: Liu Student: Ruby.
Eye Movements and Working Memory Marc Pomplun Department of Computer Science University of Massachusetts at Boston Homepage:
What factors influence movement or action? Biomechanical (e.g., size, shape, mass, strength, flexibility, coordination of body/body parts) Environmental.
What is meant by “top-down” and “bottom-up” processing? Give examples of both. Bottom up processes are evoked by the visual stimulus. Top down processes.
Chapter 5 Motor Programs 5 Motor Programs C H A P T E R.
What visual image information is needed for the things we do? How is vision used to acquire information from the world?
Lab 2 Issues: Needed to adapt to the “normal environment”. We would have liked to see more rapid adjustment and a stable baseline. Most subjects adapted.
LOGO Visual Attention in Driving: The Effects of Cognitive Load and Visual Disruption Professor: Liu Student: Ruby.
Control of Attention and Gaze in Natural Environments
CS 326A: Motion Planning Probabilistic Roadmaps for Path Planning in High-Dimensional Configuration Spaces (1996) L. Kavraki, P. Švestka, J.-C. Latombe,
Unit 3 – Driver Physical Fitness
Lecture XVII: Distributed Systems Algorithms Inspired by Biology
Issues in measuring sensory-motor control performance of human drivers: The case of cognitive load and steering control Johan Engström, Volvo Technology.
Attentional Modulations Related to Spatial Gating but Not to Allocation of Limited Resources in Primate V1  Yuzhi Chen, Eyal Seidemann  Neuron  Volume.
Wallis, JD Helen Wills Neuroscience Institute UC, Berkeley
Presentation transcript:

Control of Attention and Gaze in Natural Environments

Humans must select a limited subset of the available information in the environment. Fundamental Constraints Attention is limited. Visual Working Memory is limited. Only a limited amount of information can be retained. Evidence for this?

Selecting information from visual scenes What controls the selection process?

Image properties eg contrast, edges, chromatic saliency can account for some fixations when viewing images of scenes. Saliency - bottom-up

Limitations of Saliency Models Important information may not be salient eg Stop signs in a cluttered environment. Salient information may not be important - eg retinal image transients from eye/body movements. Doesn’t account for many observed fixations, especially in natural behavior (eg Land etc).

Natural vision is not the same as viewing pictures. Behavioral goals determine what information is needed. Task structure (often) allows interpretation of role of fixations. Need to Study Natural Behavior

Foot placement Obstacle avoidance Heading Viewing pictures of scenes is different from acting within scenes. Top-down factors

To what extent is the selection of information from scenes determined by cognitive goals (ie top-down) and how much by the stimulus itself (ie salient regions - bottom-up effects)?

Modeling Top Down Control Virtual Humanoid has a small library of simple visual behaviors: –Sidewalk Following –Picking Up Blocks –Avoiding Obstacles Each behavior uses a limited, task-relevant selection of visual information from scene. This is computationally efficient. Walter the Virtual Humanoid Sprague & Ballard (2003)

Walter learns where/when to direct gaze using reinforcement learning algorithm. Walter’s sequence of fixations obstacles sidewalk litter

Walter the Virtual Humanoid Sprague & Ballard (VSS 2004) What about unexpected events?

Dynamic Environments

Computational load Unexpected events Bottom-up ExpensiveCan handle unexpected salient events Top-down EfficientHow to deal with unexpected events?

Driving Simulator

Gaze distribution is very different for different tasks Time fixating Intersection.

The Problem Any selective perceptual system must choose the right visual computations, and when to carry them out. How do we deal with the unpredictability of the natural world? Answer - it’s not all that unpredictable and we’re really good at learning it.

Human Gaze Distribution when Walking Experimental Question: How sensitive are subjects to unexpected salient events? General Design: Subjects walked along a footpath in a virtual environment while avoiding pedestrians. Do subjects detect unexpected potential collisions?

Virtual Walking Environment Virtual Research V8 Head Mounted Display with 3 rd Tech HiBall Wide Area motion tracker V8 optics with ASL501 Video Based Eye Tracker (Left) and ASL 210 Limbus Tracker (Right) D&c emily Video Based Tracker Limbus Tracker

Virtual Environment Bird’s Eye view of the virtual walking environment. Monument

1 - Normal Walking: Avoid the pedestrians while walking at a normal pace and staying on the sidewalk. 2 - Added Task: Identical to condition 1. However, the additional instruction of following a yellow pedestrian was given Normal walking Follow leader Experimental Protocol

Pedestrians’ paths Colliding pedestrian path What Happens to Gaze in Response to an Unexpected Salient Event? The Unexpected Event: Pedestrians on a non-colliding path changed onto a collision course for 1 second (10% frequency). Change occurs during a saccade. Does a potential collision evoke a fixation?

Fixation on Collider

No Fixation During Collider Period

Probability of Fixation During Collision Period Pedestrians’ paths Colliding pedestrian path More fixations on colliders in normal walking. No effect in Leader condition Controls Colliders Normal Walking Follow Leader

Small increase in probability of fixating the collider. Failure of collider to attract attention with an added task (following) suggests that detections result from top-down monitoring. Why are colliders fixated?

Detecting a Collider Changes Fixation Strategy Longer fixation on pedestrians following a detection of a collider “Miss”“Hit” Time fixating normal pedestrians following detection of a collider Normal Walking Follow Leader

To make a top-down system work, Subjects need to learn statistics of environmental events and distribute gaze/attention based on these expectations. Subjects rely on active search to detect potentially hazardous events like collisions, rather than reacting to bottom-up, looming signals.

Possible reservations… Perhaps looming robots not similar enough to real pedestrians to evoke a bottom-up response.

Walking -Real World Experimental question: Do subjects learn to deploy gaze in response to the probability of environmental events? General design: Subjects walked on an oval path and avoided pedestrians

Experimental Setup System components: Head mounted optics (76g), Color scene camera, Modified DVCR recorder, Eye Vision Software, PC Pentium 4, 2.8GHz processor A subject wearing the ASL Mobile Eye

Occasionally some pedestrians veered on a collision course with the subject (for approx. 1 sec) 3 types of pedestrians: Trial 1: Rogue pedestrian - always collides Safe pedestrian - never collides Unpredictable pedestrian - collides 50% of time Trail 2: Rogue Safe Safe Rogue Unpredictable - remains same Experimental Design (ctd)

Fixation on Collider

Effect of Collision Probability Probability of fixating increased with higher collision probability.

Detecting Collisions: pro-active or reactive? Probability of fixating risky pedestrian similar, whether or not he/she actually collides on that trial.

Learning to Adjust Gaze Changes in fixation behavior fairly fast, happen over 4-5 encounters (Fixations on Rogue get longer, on Safe shorter)

Shorter Latencies for Rogue Fixations Rogues are fixated earlier after they appear in the field of view. This change is also rapid.

Effect of Behavioral Relevance Fixations on all pedestrians go down when pedestrians STOP instead of COLLIDING. STOPPING and COLLIDING should have comparable salience. Note the the Safe pedestrians behave identically in both conditions - only the Rogue changes behavior.

Fixation probability increases with probability of a collision. Fixation probability similar whether or not the pedestrian collides on that encounter. Changes in fixation behavior fairly rapid (fixations on Rogue get longer, and earlier, and on Safe shorter, and later)

Our Experiment: Virtual environment - want to compare real and virtual. Do observers learn to deploy visual attention based on environmental probabilities? Safe pedestrians - rarely collide Risky pedestrians - often collide Rogue pedestrians - collide a lot Do subjects fixate risky and rogue pedestrians more? How quickly does this happen?

Subjects must learn the probabilistic structure of the world and allocate gaze accordingly. That is, gaze control is model-based. Subjects behave very similarly despite unconstrained environment and absence of instructions. Control of gaze is proactive, not reactive, and thus is model based. Anticipatory use of gaze is probably necessary for much visually guided behavior. Conclusions

Behaviors Compete for Gaze/ Attentional Resources The probability of fixation is lower for both Safe and Rogue pedestrians in both the Leader conditions than in the baseline condition. Note that all pedestrians are allocated fewer fixations, even the Safe ones.

Conclusions Data consistent with task-driven sampling of visual information rather than bottom up capture of attention - No effect of increased salience of collision event. - Colliders fail to attract gaze in the leader condition, suggesting the extra task interferes with detection. Observers rapidly learn to deploy visual attention based on environmental probabilities. Such learning is necessary in order to deploy gaze and attention effectively.

Certain stimuli thought to capture attention bottom-up (eg Theeuwes et al, 2001 etc ) Looming stimuli seem like good candidates for bottom-up attentional capture (Regan & Gray, 200; Franceroni & Simons,2003).

No Leader Normal WalkingFollow Leader Greater saliency of the unexpected event does not increase fixations. No effect of increased collider speed.

Other evidence for detection of colliders? Do subjects slow down during collider period? Subjects slow down, but only when they fixate collider. Implies fixation measures “detection”. Slowing is greater if not previously fixated. Consistent with peripheral monitoring of previously fixated pedestrians.

Conclusions Subjects learn the probabilities of events in the environment and distribute gaze accordingly The findings from the Leader manipulation support the claim that different tasks compete for attention

Effect of Context Probability of fixating Safe pedestrian higher in a context of a riskier environment

Summary Direct comparison between real and virtual collisions is difficult, but colliders are still not reliably fixated. Subjects appear to be sensitive to several parameters of the environment: –Experience Experience with the Rogue pedestrian elevated fixation probabilities of the Safe pedestrian to 70% (50% wto. exp.) Experience with the Safe lead to 80% fixation probability of the Rogue (89% wto. exp.) Experience of Safe carries less weight than the experience of Rogue

Time fixating Intersection. “Follow the car.” or “Follow the car and obey traffic rules.” CarRoadsideRoadIntersection Shinoda et al. (2001) Detection of signs at intersection results from frequent looks.

Intersection P = 1.0 Mid-block P = 0.3 Greater probability of detection in probable locations Suggests Ss learn where to attend/look. How well do human subjects detect unexpected events? Shinoda et al. (2001) Detection of briefly presented Stop signs.

What do Humans Do? Shinoda et al. (2001) found better detection of unexpected stop signs in a virtual driving task.