Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000.

Slides:



Advertisements
Similar presentations
Université du Québec École de technologie supérieure Face Recognition in Video Using What- and-Where Fusion Neural Network Mamoudou Barry and Eric Granger.
Advertisements

Active Appearance Models
Face Recognition: A Convolutional Neural Network Approach
Introduction to Cognitive Science Adrienne Moore, Office hours: Wed. 4-5, Cognitive Science Building,
(2) Face Recognition These notes are the second part of a two-part lecture roughly corresponding to (1) object recognition and (2) face recognition We'll.
Chapter 4: Cortical Organization
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
Comparing Thompson’s Thatcher effect with faces and non-face objects Elyssa Twedt 1, David Sheinberg 2 & Isabel Gauthier 1 Vanderbilt University 1, Brown.
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
BEYOND SIMPLE FEATURES: A LARGE-SCALE FEATURE SEARCH APPROACH TO UNCONSTRAINED FACE RECOGNITION Nicolas Pinto Massachusetts Institute of Technology David.
On the Relationship between Visual Attributes and Convolutional Networks Paper ID - 52.
Face Recognition Under Varying Illumination Erald VUÇINI Vienna University of Technology Muhittin GÖKMEN Istanbul Technical University Eduard GRÖLLER Vienna.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.
Robust Real-time Object Detection by Paul Viola and Michael Jones ICCV 2001 Workshop on Statistical and Computation Theories of Vision Presentation by.
Dynamic Face Recognition Committee Machine Presented by Sunny Tang.
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
October 14, 2010Neural Networks Lecture 12: Backpropagation Examples 1 Example I: Predicting the Weather We decide (or experimentally determine) to use.
Spatial Pyramid Pooling in Deep Convolutional
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Oral Defense by Sunny Tang 15 Aug 2003
CS 484 – Artificial Intelligence
Radial-Basis Function Networks
Facial Recognition CSE 391 Kris Lord.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Database Construction for Speech to Lip-readable Animation Conversion Gyorgy Takacs, Attila Tihanyi, Tamas Bardi, Gergo Feldhoffer, Balint Srancsik Peter.
November 25, 2014Computer Vision Lecture 20: Object Recognition IV 1 Creating Data Representations The problem with some data representations is that the.
West Virginia University
Thien Anh Dinh1, Tomi Silander1, Bolan Su1, Tianxia Gong
Solving “The Visual Expertise Mystery” with Models Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) The Perceptual Expertise Network The Temporal.
Presented by Tienwei Tsai July, 2005
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Biases: An Example Non-accidental properties: Properties that appear in an image that are very unlikely to have been produced by chance, and therefore.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Rotation Invariant Neural-Network Based Face Detection
Local Non-Negative Matrix Factorization as a Visual Representation Tao Feng, Stan Z. Li, Heung-Yeung Shum, HongJiang Zhang 2002 IEEE Presenter : 張庭豪.
ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and.
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
A Face processing system Based on Committee Machine: The Approach and Experimental Results Presented by: Harvest Jang 29 Jan 2003.
EE459 Neural Networks Examples of using Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
Talk #3: EMPATH A Neural Network Model of Human Facial Expression Recognition Gary Cottrell Joint work with Matt Dailey, Curt Padgett, and Ralph Adolphs.
Feature selection with Neural Networks Dmitrij Lagutin, T Variable Selection for Regression
EMPATH: A Neural Network that Categorizes Facial Expressions Matthew N. Dailey and Garrison W. Cottrell University of California, San Diego Curtis Padgett.
Face Detection Using Neural Network By Kamaljeet Verma ( ) Akshay Ukey ( )
2D-LDA: A statistical linear discriminant analysis for image matrix
Chapter 4: Cortical Organization
Meeting 8: Features for Object Classification Ullman et al.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
MicroArray Data Analysis Candice Quadros & Amol Kothari.
Do Expression and Identity Need Separate Representations?
Artificial Neural Networks
Compact Bilinear Pooling
CLPS0020: Introduction to Cognitive Science
Radial Basis Function G.Anuradha.
Ajita Rattani and Reza Derakhshani,
Recovery from Occlusion in Deep Feature Space for Face Recognition
Context-based vision system for place and object recognition
Face Recognition and Detection Using Eigenfaces
CS 4501: Introduction to Computer Vision Training Neural Networks II
Zhengjun Pan and Hamid Bolouri Department of Computer Science
Creating Data Representations
Face Recognition: A Convolutional Neural Network Approach
Do Better ImageNet Models Transfer Better?
Presentation transcript:

Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000

Overview Is the Fusiform Face Area (FFA) really a face specific area? Our results support the view that it is NOT Motivation : Evidence for/against the face specific view Our Approach : Our model and experimental design Results Conclusion

Motivation: Evidence for the Face Specific View Prosopagnosia patients may have deficit in identifying individual faces but normal in detecting faces or other non-face objects, while visual object agnosia patients may be normal with face recognition but have deficit with reading or object recognition. Recognition of faces is more sensitive to configural changes than objects. Face and non-face objects have separate processing mechanisms

Motivation: Evidence against the face specific view Gauthier et al. points out faces and objects differ not only in the image geometries, but also in … 1.Level of discrimination 2.Level of experience We are face “experts”. FFA showed high activation for a wide variety of non-objects when these two conditions were controlled.

Greeble Experts (Gauthier et al. 1999) Activation of the FFA increased when Greebles were presented as the training proceeded. When subjects met the criteria of experts, the activation level differences between faces and Greebles were insignificant.

Our Hypothesis Why does the FFA engage in expert classification of non-face objects as well? We hypothesized that the FFA responds to visual features that are generally useful in discriminating homogeneous input images. Expertise on one class should facilitate the learning of other expert tasks.

Model Pretrain two groups of neural networks on different tasks. Compare the abilities to learn a new individual Greeble classification task. cup Carol book can Bob Ted can cup book face Hidden layer Greeble1 Greeble2 Greeble3 Greeble1 Greeble2 Greeble3 (Experts) (Non-experts)

Database 64x64 8bit grayscale 5 basic categories 12 individuals per category 5 different images per individual Total of 5x12x5=300 images

Preprocessing Form Gabor jets using 8 orientations and 5 scales, then subsample on a 8x8 grid. For each scale, apply PCA separately and reduce dimensionality to 8. 8x5x64=2560 5x8=40 dimensions

Experimental Setting Details Fixed configurations: 1 hidden layer with 40 units Learning rate =.005 Momentum =.5 Controlled condition: 30 training patterns for each basic class can cup book face can cup book face1 face2 face3 face10 (Experts) (Non-experts) Pretraining tasks

Training Set Variations Training set: 10 individual for each class, 3 images for each individual Hold out / test set: Basic level: 3 images of unseen individual Individual level: 1 unseen image of each individual Training set Hold out Test (indiv.) 10 Hold out Test (basic)

Topics for Analysis Will the experts learn the new task faster? Is there a correlation between network plasticity and the speed to learn the new task? Network plasticity can be defined as the average “slope” of the hidden layer units.

Criteria to Stop Training 1.Fixed RMSE threshold 2.Number of training epochs 3.Best holdout set performance We eliminated the third criterion due to extremely high variance in number of training epochs observed in preliminary studies.

Experiment 1 (Design) For each pretraining task, 20 networks were trained, i.e. 20 book experts, 20 face experts, etc. Training set RMSE threshold was fixed to: –.08 (pretraining) –.158 (after new task added) The threholds were derived from preliminary cross-validation experiments on the most difficult task, i.e. face expert classification.

Experiment 1 (Results) Pretraining tasks were much harder for the experts. Non-experts were significantly slower in learning the new task than any of the experts. Number of Epochs to Achieve RMSE Threshold

Experiment 2 (Design) For each pretraining task, 10 networks for trained for 5120 epochs. Intermediate weights were recorded at epochs 5,10, 20,..., Total of 11x10=110 networks were trained on the new task with RMSE threshold.158.

Experiment 2 (Results) Non-experts were the slowest to learn the new task, provided that the pretraining tasks were fully acquired. Pretraining Epochs and the Speed to Learn New Task

Experiment 2 (Results) If the pretraining were stopped prematurely, the networks must continue improving on the pretrained classes as well as the new task. Pretraining RMSE

Analysis of Network Plasticity Network plasticity can be defined as the average “slope” across all hidden layer units and all patterns in a given set of patterns: where

Network Plasticity Plasticity was lower for the experts! It is more appropriate to interpret plasticity as a measurement of mismatch.

Conclusion Expert networks learned the new task faster. Expert networks learned faster despite the low plasticity, further supporting the claim that the hidden layer representation developed for one expert class are general features that are useful for classifying other classes as well. Visual expertise is a general skill that is not specific to any class of images including faces.