Dense Regions of View-Invariant Features Promote Object Recognition

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Advertisements

Author :Andrea Selinger Salgian Department of Computer Science
The Role of Competition in Repetition Blindness Mary L. Still Alison L. MorrisIowa State University The Role of Competition in Repetition Blindness Mary.
Presented by Xinyu Chang
Evaluating Color Descriptors for Object and Scene Recognition Koen E.A. van de Sande, Student Member, IEEE, Theo Gevers, Member, IEEE, and Cees G.M. Snoek,
Effect of Opacity of Stimulus in Deployment of Interest in an Interface Sujoy Kumar Chowdhury & Jeremiah D. Still Missouri Western State University Introduction.
LOGO The role of attentional breadth in perceptual change detection Professor: Liu Student: Ruby.
›SIFT features [3] were computed for 100 images (from ImageNet [4]) for each of our 48 subordinate-level categories. ›The visual features of an image were.
Semantic consistency versus perceptual salience in visual scenes: Findings from change detection Spotorno, S., Tatler, B. W., & Faure, S. (2013). Acta.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Crowd Size Estimation Luis Huang ECE 172A - UCSD.
A Study of Approaches for Object Recognition
ACKNOWLEDGMENTS REFERENCES -Are view-boundaries special – eliciting extrapolation? There are other boundaries within a view that have similar characteristics,
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
Multiscale Moment-Based Painterly Rendering Diego Nehab and Luiz Velho
Studying Memory Encoding with fMRI Event-related vs. Blocked Designs Aneta Kielar.
The Effects of Focused Attention and Varied Peripheral and Central Changes on Change Blindness and Change Detection Teal Maxwell Emily Welch Naomi Janett.
Evaluation of interest points and descriptors. Introduction Quantitative evaluation of interest point detectors –points / regions at the same relative.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Puzzle Solver Sravan Bhagavatula EE 638 Project Stanford ECE.
Department of Psychology & The Human Computer Interaction Program Vision Sciences Society’s Annual Meeting, Sarasota, FL May 13, 2007 Jeremiah D. Still,
Region-Based Saliency Detection and Its Application in Object Recognition IEEE TRANSACTIONS ON CIRCUITS AND SYSTEM FOR VIDEO TECHNOLOGY, VOL. 24 NO. 5,
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
A Review of “Iconic Memory Requires Attention” by Persuh, Genzer, & Melara (2012) By Richard Thripp EXP 6506 – University of Central Florida September.
An Eyetracking Analysis of the Effect of Prior Comparison on Analogical Mapping Catherine A. Clement, Eastern Kentucky University Carrie Harris, Tara Weatherholt,
LOGO Change blindness in the absence of a visual disruption Professor: Liu Student: Ruby.
Disrupting face biases in visual attention Anna S. Law, Liverpool John Moores University Stephen R. H. Langton, University of Stirling Introduction Method.
Image features and properties. Image content representation The simplest representation of an image pattern is to list image pixels, one after the other.
Distinctive Image Features from Scale-Invariant Keypoints Presenter :JIA-HONG,DONG Advisor : Yen- Ting, Chen 1 David G. Lowe International Journal of Computer.
WLD: A Robust Local Image Descriptor Jie Chen, Shiguang Shan, Chu He, Guoying Zhao, Matti Pietikäinen, Xilin Chen, Wen Gao 报告人:蒲薇榄.
Body Position Influences Maintenance of Objects in Visual Short-Term Memory Mia J. Branson, Joshua D. Cosman, and Shaun P. Vecera Department of Psychology,
Feature Binding: Not Quite So Pre-attentive Erin Buchanan and M
David Marchant, Evelyn Carnegie, Paul Ellison
Computational Vision --- a window to our brain
Lab Roles and Lab Report
PERCEIVED BEAUTY OF RANDOM DENSITY PATTERNS
Human factors issues with the use of text-speak communication
Developing Group Stereotypes Over Time
Video Google: Text Retrieval Approach to Object Matching in Videos
Seunghui Cha1, Wookhyun Kim1
Perception Unit How can we so easily perceive the world? Objects may be moving, partially hidden, varying in orientation, and projected as a 2D image.
Angry Faces Capture Attention But Do They Hold It?
Enhanced-alignment Measure for Binary Foreground Map Evaluation
Brain States: Top-Down Influences in Sensory Processing
Volume 61, Issue 5, Pages (March 2009)
Two randomised controlled crossover studies to evaluate the effect of colouring on both self-report and performance measures of well-being Holt, N. J.,
Investigating the Attentional Blink With Predicted Targets
47th Annual Meeting of the Psychonomic Society, Houston, TX
Volume 88, Issue 3, Pages (November 2015)
Perceptual Echoes at 10 Hz in the Human Brain
Daphna Shohamy, Anthony D. Wagner  Neuron 
KFC: Keypoints, Features and Correspondences
Adaboost for faces. Material
Alteration of Visual Perception prior to Microsaccades
Volume 93, Issue 2, Pages (January 2017)
Jason Samaha, Bradley R. Postle  Current Biology 
Volume 79, Issue 4, Pages (August 2013)
Consolidation Promotes the Emergence of Representational Overlap in the Hippocampus and Medial Prefrontal Cortex  Alexa Tompary, Lila Davachi  Neuron 
Karl R Gegenfurtner, Jochem Rieger  Current Biology 
Machine Learning for Visual Scene Classification with EEG Data
Consequences of the Oculomotor Cycle for the Dynamics of Perception
Brain States: Top-Down Influences in Sensory Processing
Video Google: Text Retrieval Approach to Object Matching in Videos
Decoding the Yellow of a Gray Banana
Consequences of the Oculomotor Cycle for the Dynamics of Perception
Computer Vision in Cell Biology
Multi-Information Based GCPs Selection Method
Beth L. Parkin, Hamed Ekhtiari, Vincent F. Walsh  Neuron 
Volume 20, Issue 7, Pages (April 2010)
Perceptual-Motor Deficits in Children with down syndrome: Implications for Intervention Study by: Naznin Virji-Babul, Kimberly Kerns, Eric Zhou, Asha.
Presentation transcript:

Dense Regions of View-Invariant Features Promote Object Recognition Object Perception, Attention, & Memory 16th Annual Meeting, Chicago, Illinois November 13, 2008 Human-Computer Interaction Program & Department of Psychology Jeremiah D. Still, Veronica J. Dark, & Derrick J. Parkhurst Dense Regions of View-Invariant Features Promote Object Recognition Introduction Although saliency (Parkhurst, Law & Niebur, 2002) and invariance (as determined by Scale Invariant Feature Transform – SIFT; Lowe, 1999) are highly correlated, we found that invariance more accurately predicted overt attentional selection, as measured by eye movements, than did salience (Still, Dark & Parkhurst, 2007). To investigate whether invariant features that are important for computer vision really contribute to human object recognition, we created two photograph fragments for each photo. One condition contained 50% of the pixels associated with the densest regions of invariant features as defined by the adapted SIFT algorithm (see Figure 2). The other condition, the complement, contained the remaining 50% of the pixels. We found that objects were more easily identified when the fragments contained more invariant features (Wolff, Still, Parkhurst & Dark, 2007). In the current experiment, we manipulated the percentage of pixels displayed with the displayed regions selected either randomly or in terms of density of invariant features. If invariant features are critical for object recognition, naming accuracy should be higher for the invariant conditions compared to the random conditions. In addition, the benefit for invariant regions should decrease as number of pixels is increased. Method Stimuli were a set of 64 nameable photographs of complex objects from the Amsterdam Library of Object Images (Geusebroek, Burghouts and Smeulders, 2005). Eight fragments were constructed for each object. The fragments varied in whether pixel regions were chosen randomly or in terms of density of invariant features (see Figure 1); they also varied in the proportion of pixels chosen (20%, 40%, 60%, 80%; See Figure 3). Figure 1: Creating an Invariant Features Pre-attentional Map Results Figure 2: Schematic of our Adaptation of the SIFT Algorithm Figure 4: Participant Object Recognition Naming Accuracy Standard Error Bars Invariant fragments (i.e., those constructed to capture invariant features) were more likely to be named correctly than random fragments F(1, 31) = 24.88, MSE = .016, p < .001. As the proportion of fragments increased, naming accuracy increased, F(3, 29) = 43.91, MSE = .027, p <.001. We had predicted that the benefit to invariant fragments would decrease as the proportion of pixels shown increased. Although the predicted interaction was not significant, one-tailed comparisons showed that accuracy in the invariant condition was higher than in the random condition with 20% pixels, t(31) = 2.10, SEM = .04, p = .02, and with 40% pixels, t(31) = 2.44, SEM = .05, p = .01. The difference was marginally significant with 60% pixels, t(31) = 1.66, SEM = .04, p = .05, but not significant with 80% pixels, t < 1, p = .26. Discussion Regions containing invariant features appear to be important for human object perception, confirming our previous findings (Wolff et al., 2007). The same type of invariance that is important for computer vision (Lowe, 1999) was important for human object recognition. It is possible that dense regions of invariant features are extracted by a pre-attentive, low-level system and that these regions provide information for programming eye movements (Rensink, 2000). This guidance of attention to invariant regions may play an early role in object perception. Taken together, our research supports the hypothesis that the visual system is biased to select visual features likely to be important for object recognition. Procedure Participants (N = 32) viewed each fragmented object for 1,000 msec and then typed the object name, guessing when unsure. Fragment type (invariant or random) and proportion of pixels were manipulated within-subjects with each participant viewing 8 images for each of the 8 conditions. Counterbalancing insured that each participant saw only one version of an object and that each object was presented in all conditions. Figure 3: Example Stimuli References Geusebroek, J., Burghouts, G., J., & Smeulders, A., W. M. (2005). The Amsterdam library of object images. International Journal of Computer Vision, 61(1), 103-112. Lowe, D. G. (1999, September). Object recognition from local scale-invariant features. Paper presented at the International Conference on Computer Vision, Corfu, Greece, 1150-1157. Lowe, D.G. (2004). Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60(2), 91-110. Parkhurst, D. J., Law, K., & Niebur, E. (2002). Modeling the role of salience in the allocation of overt visual attention. Vision Research, 42, 107-123. Rensink, R. A. (2000). The dynamic representation of scenes. Visual Cognition, 7, 17-42. Still, J. D., Dark, V. J., & Parkhurst, D. J. (2007, May). Viewpoint Invariant Object Features Attract Overt Visual Attention. Poster presented at the meeting of the Vision Sciences Society, Sarasota, FL. Wolff, T., Still, J. D., Parkhurst, D. J. & Dark, V. J. (2007, May). Invariant Features Detected with Computer Vision Allow Better Human Object Recognition in Photographs. Poster presented at the meeting of the Midwestern Psychological Association, Chicago, IL. JeremiahStill.info