Saliency & attention (P) Lavanya Sharan April 4th, 2011.

Slides:



Advertisements
Similar presentations
Caroline Rougier, Jean Meunier, Alain St-Arnaud, and Jacqueline Rousseau IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 21, NO. 5,
Advertisements

Interest points CSE P 576 Ali Farhadi Many slides from Steve Seitz, Larry Zitnick.
Detail to attention: Exploiting Visual Tasks for Selective Rendering Kirsten Cater 1, Alan Chalmers 1 and Greg Ward 2 1 University of Bristol, UK 2 Anyhere.
Different Pathways, Different Processes. Retinocollicular vs. Retinostriate Recall that 10% of optic nerve gets routed through the Superior Colliculus.
Wrap Up. We talked about Filters Edges Corners Interest Points Descriptors Image Stitching Stereo SFM.
LOGO The role of attentional breadth in perceptual change detection Professor: Liu Student: Ruby.
Chapter 6: Visual Attention. Scanning a Scene Visual scanning – looking from place to place –Fixation –Saccadic eye movement Overt attention involves.
LOGO Effects of scene inversion on change detection of targets matched for visual salience Professor: Liu Student: Ruby.
RESEARCH METHODS Lecture 18
Knowing a Good HOG Filter When You See It: Efficient Selection of Filters for Detection Ejaz Ahmed 1, Gregory Shakhnarovich 2, and Subhransu Maji 3 1 University.
Control of Attention and Gaze in the Natural World.
Image Retrieval Using Eye Movements Fred Stentiford & Wole Oyekoya University College London.
Read this paper Chellazi et al. (1993) Nature 363 Pg
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Quantifying and Transferring Contextual Information in Object Detection Professor: S. J. Wang Student : Y. S. Wang 1.
Interest points CSE P 576 Larry Zitnick Many slides courtesy of Steve Seitz.
Genetic Algorithms  An example real-world application.
Eye Tracking for Personalized Photography Steven Scher (UCSC student) James Davis (UCSC advisor) Sriram Swaminarayan (LANL advisor)
Visual Attention More information in visual field than we can process at a given moment Solutions Shifts of Visual Attention related to eye movements Some.
Tracking multiple independent targets: Evidence for a parallel tracking mechanism Zenon Pylyshyn and Ron Storm presented by Nick Howe.
Attention. Looking without Seeing.
1 Outline Overview Integrating Vision Models CCM: Cascaded Classification Models Learning Spatial Context TAS: Things and Stuff Descriptive Querying of.
What information are people “seeing” when your web page pops up? By Shirley Retz.
Overview Introduction to local features
Sung Ju Hwang and Kristen Grauman University of Texas at Austin.
Sung Ju Hwang and Kristen Grauman University of Texas at Austin CVPR 2010.
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
Exploiting video information for Meeting Structuring ….
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Vision Surveillance Paul Scovanner.
Dynamic 3D Scene Analysis from a Moving Vehicle Young Ki Baik (CV Lab.) (Wed)
Model comparison and challenges II Compositional bias of salient object detection benchmarking Xiaodi Hou K-Lab, Computation and Neural Systems California.
Visual Attention Derek Hoiem March 14, 2007 Misc Reading Group.
Assessment of Computational Visual Attention Models on Medical Images Varun Jampani 1, Ujjwal 1, Jayanthi Sivaswamy 1 and Vivek Vaidya 2 1 CVIT, IIIT Hyderabad,
Essence The of the Scene Gist Trayambaka Karra KT and Garold Fuks.
Why is computer vision difficult?
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
Learning Analytics: Process & Theory March 24, 2014.
Look Over Here: Attention-Directing Composition of Manga Elements Ying Cao Rynson W.H. Lau Antoni B. Chan SIGGRAPH
Automated Reading Assistance System Using Point-of-Gaze Estimation M.A.Sc. Thesis Presentation Automated Reading Assistance System Using Point-of-Gaze.
1 Self-Calibration and Neural Network Implementation of Photometric Stereo Yuji IWAHORI, Yumi WATANABE, Robert J. WOODHAM and Akira IWATA.
Overview Introduction to local features Harris interest points + SSD, ZNCC, SIFT Scale & affine invariant interest point detectors Evaluation and comparison.
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Older Driver Failures of Attention at Intersections: Using Change Blindness Methods to Assess Turn Decision Accuracy Professor: Liu Student: Ruby.
 Example: seeing a bird that is singing in a tree or miss a road sign in plain sight  Cell phone use while driving reduces attention and memory for.
Picture change during blinks: looking without seeing and seeing without looking Professor: Liu Student: Ruby.
Optimal Eye Movement Strategies In Visual Search.
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Using the Forest to see the Trees: A computational model relating features, objects and scenes Antonio Torralba CSAIL-MIT Joint work with Aude Oliva, Kevin.
Describe how reaching and grasping abilities develop in the first year of life.
Demonstration Study for Applying AVED to Still Images from Station M Update, next steps, workflow overview Demonstration Study for Applying AVED to Still.
A computational model of stereoscopic 3D visual saliency School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
Max-Margin Training of Upstream Scene Understanding Models Jun Zhu Carnegie Mellon University Joint work with Li-Jia Li *, Li Fei-Fei *, and Eric P. Xing.
Learning Hierarchical Features for Scene Labeling Cle’ment Farabet, Camille Couprie, Laurent Najman, and Yann LeCun by Dong Nie.
Real-Time Hierarchical Scene Segmentation and Classification Andre Uckermann, Christof Elbrechter, Robert Haschke and Helge Ritter John Grossmann.
ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
Journal of Vision. 2016;16(3):39. doi: / Figure Legend:
From: What's color got to do with it
Saliency detection Donghun Yeo CV Lab..
Context-based vision system for place and object recognition
Object detection as supervised classification
Crowding by a single bar
RESEARCH METHODS Lecture 18
Volume 88, Issue 3, Pages (November 2015)
Dense Regions of View-Invariant Features Promote Object Recognition
Structure of a typical back-propagated multilayered perceptron used in this study. Structure of a typical back-propagated multilayered perceptron used.
Bland-Altman plots showing intraobserver variability of sTBF measurements for observers 1 (A and C) and 2 (B and D) with maximum (A and B) and average.
James P. Howard et al. JACEP 2019;j.jacep
Presentation transcript:

Saliency & attention (P) Lavanya Sharan April 4th, 2011

Looking for people Ehinger et al. (2009) Question: Where do people look and can we model it?

Looking for people Ehinger et al. (2009) 14 observers with tracking accuracy 0.75 deg. Observers asked to look for pedestrians as fast as possible and answer. 912 images. Target (pedestrian) present in half the images. Images 23.5 x 17.7 deg, target when present 0.9 x 1.8 deg.

Inter-observer agreement Ehinger et al. (2009) Question: Do observers agree with each other? Use fixations from n-1 observers for one image, apply gaussian blur and use this region to predict fixation of remaining observer on that image. Repeat for n observers (Torralba et al. 2006). Do this for all images.

Inter-observer agreement Ehinger et al. (2009)

Cross-image control Ehinger et al. (2009) Question: Do fixated locations depend on image content? Or do observers always look at the same places? Use fixations from n-1 observers for one image, apply gaussian blur and use this region to predict fixation of remaining observer on a different, randomly selected image. Repeat for n observers (Torralba et al. 2006). Do this for all images.

Performance at predicting fixation locations Ehinger et al. (2009) Inter-observer agreement (Target absent AUC=0.93, target present AUC = 0.95) Cross-image control (Target absent AUC=0.68, target present AUC = 0.62)

Modeling eye movements during pedestrian detection Ehinger et al. (2009)

Low-level saliency. Use Torralba et al model. Target features. Use Dalal et al person detector. Scene context. Use Torralba et al. model but train on context (i.e. sidewalks not skies are good locations) + ‘Context Oracle’ (7 observers indicate good candidates for locations) Use validation set to decide relative weighting.

Modeling eye movements during pedestrian detection Ehinger et al. (2009)

Performance of saliency component

Modeling eye movements during pedestrian detection Ehinger et al. (2009) Performance of target features component

Modeling eye movements during pedestrian detection Ehinger et al. (2009) Performance of scene context component

Modeling eye movements during pedestrian detection Ehinger et al. (2009) Performance of full model

What does the model get wrong? Ehinger et al. (2009)