Do Expression and Identity Need Separate Representations?

Slides:



Advertisements
Similar presentations
FACIAL EMOTION RECOGNITION BY ADAPTIVE PROCESSING OF TREE STRUCTURES Jia-Jun Wong and Siu-Yeung Cho Forensic and Security Lab School of Computer Engineering.
Advertisements

Matthias Wimmer, Ursula Zucker and Bernd Radig Chair for Image Understanding Computer Science Technische Universität München { wimmerm, zucker, radig
Age Differences in Emotion Recognition of Briefly Presented Faces Lisa Emery, Kory Morgan, Kaitlyn Pechanek & Caitlin Williams Reprints may be obtained.
Robust 3D Head Pose Classification using Wavelets by Mukesh C. Motwani Dr. Frederick C. Harris, Jr., Thesis Advisor December 5 th, 2002 A thesis submitted.
HYBRID-BOOST LEARNING FOR MULTI-POSE FACE DETECTION AND FACIAL EXPRESSION RECOGNITION Hsiuao-Ying ChenChung-Lin Huang Chih-Ming Fu Pattern Recognition,
I. Face Perception II. Visual Imagery. Is Face Recognition Special? Arguments have been made for both functional and neuroanatomical specialization for.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Un Supervised Learning & Self Organizing Maps Learning From Examples
CONTENT BASED FACE RECOGNITION Ankur Jain 01D05007 Pranshu Sharma Prashant Baronia 01D05005 Swapnil Zarekar 01D05001 Under the guidance of Prof.
Feature Detection and Emotion Recognition Chris Matthews Advisor: Prof. Cotter.
TAUCHI – Tampere Unit for Computer-Human Interaction Automated recognition of facial expressi ns and identity 2003 UCIT Progress Report Ioulia Guizatdinova.
Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Introduction to Wavelets
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
EKMAN’S FACIAL EXPRESSIONS STUDY A Demonstration.
Emotions: Emotions: Hampson, E., van Anders, S. M., & Mullin, L. I. (2006). A female advantage in the recognition of emotional facial expressions: test.
Recognizing Emotions in Facial Expressions
Face Recognition CPSC 601 Biometric Course.
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Face Recognition Using EigenFaces Presentation by: Zia Ahmed Shaikh (P/IT/2K15/07) Authors: Matthew A. Turk and Alex P. Pentland Vision and Modeling Group,
Facial Feature Detection
SSIP Project 2 GRIM GRINS Michal Hradis Ágoston Róth Sándor Szabó Ilona Jedyk Team 2.
West Virginia University
Human Emotion Synthesis David Oziem, Lisa Gralewski, Neill Campbell, Colin Dalton, David Gibson, Barry Thomas University of Bristol, Motion Ripper, 3CR.
1 Facial Expression Recognition using KCCA with Combining Correlation Kernels and Kansei Information Yo Horikawa Kagawa University, Japan.
1 Backprop, 25 years later… Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science and Engineering Department Temporal Dynamics.
Multimodal Information Analysis for Emotion Recognition
EE459 Neural Networks Examples of using Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 25 Nov 4, 2005 Nanjing University of Science & Technology.
Talk #3: EMPATH A Neural Network Model of Human Facial Expression Recognition Gary Cottrell Joint work with Matt Dailey, Curt Padgett, and Ralph Adolphs.
Intelligent Control and Automation, WCICA 2008.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
EMPATH: A Neural Network that Categorizes Facial Expressions Matthew N. Dailey and Garrison W. Cottrell University of California, San Diego Curtis Padgett.
3D Face Recognition Using Range Images
Introduction Performance of metric learning is heavily dependent on features extracted Sensitive to Performance of Filters used Need to be robust to changes.
Cognitive models for emotion recognition: Big Data and Deep Learning
Face Recognition Summary –Single pose –Multiple pose –Principal components analysis –Model-based recognition –Neural Networks.
Ekman’s Facial Expressions Study A Demonstration.
Facial Expressions and Emotions Mental Health. Total Participants Adults (30+ years old)328 Adults (30+ years old) Adolescents (13-19 years old)118 Adolescents.
Evaluation of Gender Classification Methods with Automatically Detected and Aligned Faces Speaker: Po-Kai Shen Advisor: Tsai-Rong Chang Date: 2010/6/14.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
Emotion Knowledge in Maltreated Preschoolers
British face stimuli Egyptian face stimuli
Face Recognition using Artificial Neural Network
EMOTIONAL INTELLIGENCE
Angry Faces Capture Attention But Do They Hold It?
Feature based vs. holistic processing
Final Year Project Presentation --- Magic Paint Face
Gary Cottrell & * 07/16/96 The Face of Fear: A Neural Network Model of Human Facial Expression Recognition Garrison W. Cottrell Gary's Unbelievable Research.
Feature based vs. holistic processing
Crowding by a single bar
Domingo Mery Department of Computer Science
Emotions cse 574 winter 2004.
network of simple neuron-like computing elements
Cognitive Processes PSY 334
Backprop, Representations, Representations, Representations,
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Cultural Confusions Show that Facial Expressions Are Not Universal
Itzhak Fried, Katherine A MacDonald, Charles L Wilson  Neuron 
Sparseness and Expansion in Sensory Representations
Uma R. Karmarkar, Dean V. Buonomano  Neuron 
Attention for translation
Encoding of Stimulus Probability in Macaque Inferior Temporal Cortex
by Khaled Nasr, Pooja Viswanathan, and Andreas Nieder
Bruce & Young’s model of face recognition (1986)
Presentation transcript:

Do Expression and Identity Need Separate Representations? Kristin Branson (kbranson@cs.ucsd.edu), Gary Cottrell (gary@cs.ucsd.edu), Andrew Calder (andy.calder@mrc-cbu.cam.ac.uk) 24th Annual Meeting of the Cognitive Science Society Friday, August 9, 2002

Introduction Facial identity recognition uses a holistic, or configural representation [Tanaka & Farah, 1993], [Young et al., 1987]. A configural model incorporates information about the relationship between (first-order) features. Does facial expression recognition use a configural representation? Are the representations for expression and identity separate? Example First-Order Features Example Configural Information

Calder’s Behavioral Experiments Andrew Calder performed experiments to answer these questions, with the conclusions: Expression recognition uses a configural model. The representations for expression and identity are separate [Calder, et al., 2000]. Identity Representation Expression Task Cathy Happy

Purpose Our goal is to model Calder’s experiments using the Dailey et al. model of expression recognition, and a single representation for expression and identity. If we obtain similar results, then Calder’s results can be obtained using only one representation for identity and expression. Face Representation Neural Network Cathy Happy

Stimuli: Composite Images Two types of stimuli: (Aligned) Composite and (Misaligned) Noncomposite face images. (Aligned) Composite images are created by aligning e.g. the top half of a fearful face with the bottom half of a disgusted face. (Aligned) Composite “nr” posing “fearful” “nr” posing “disgusted”

Stimuli: Noncomposite Images (Misaligned) Noncomposite images are created by misaligning e.g. the top half of a fearful face with the bottom half of a disgusted face. (Misaligned) Noncomposite “nr” posing “fearful” “nr” posing “disgusted”

Evidence for a Configural Model Incorrect configural information from the other half of the face is present in composite but not noncomposite images. If it is harder to identify the expression in half a composite than a noncomposite, incorrect configural information disrupts expression recognition, demonstrating that a configural model is used. This method parallels experiments performed to demonstrate that a configural model is used for identity recognition [Young, et al., 1987]. Incorrect Configural Information Composite Noncomposite

Supporting Independence Incorrect identity configural information Incorrect expression configural information Same Expression, Different Identity Different Expression, Same Identity Incorrect identity and expression configural information Different Expression, Different Identity Create three different types of composite images.

Supporting Independence SID, DE DID, DE DID, SE (SID,DE and DID,DE) contain incorrect expression configural information. Expression recognition is disrupted. If expression recognition is independent of identity recognition, incorrect identity configural information should not affect expression recognition. Therefore, it should be no harder to recognize expression in the DID, DE than the SID, DE composites. (DID,SE and DID,DE) contain incorrect identity configural information. Identity recognition is disrupted. If identity recognition is independent of expression recognition, incorrect expression configural information should not affect identity recognition. Therefore, it should be no harder to recognize identity in the DID, DE than the DID, SE composites.

Stimuli Preprocessing Gabor Wavelet Filtering The pixel image is convolved with 2D Gabor wavelet filters of 8 different orientations and 5 different scales. Results in a 40,600 element vector. Similar to filtering in the striate cortex of cats; reacts strongly to edges. Insensitive to small translations. 2D Gabor wavelet filters of different scales and orientations [Daugman, 1985] Images of 10 actors posing 6 expressions from the Ekman & Friesen database of Pictures of Facial Affects. Six “universal” expressions (Happy, Sad, Angry, Fearful Surprised, Disgusted). Actors are trained to move some of 44 muscle groups identified by Ekman as displaying the posed expression. All expressions were recognizable by at least 70% of those tested on them. Pixel Image Database Gabor Pattern Gabor Filtering Attenuated Pattern Attenuate If the current experiment requires that the network identify the expression in only one half of the face, attention to the other half of the face must be attenuated. This is done by multiplying the elements of the Gabor pattern in the half we are attenuating by a fraction. Input Pixel Image

Single-layer, feed-forward neural net Model Description Happy Sad PCA Angry Preprocessing Fearful Surprised Pixel Image Disgusted Network Output Single-layer, feed-forward neural net

Model Training Training stimuli in all experiments: The model is trained on nine actors and tested on the tenth. This is repeated for all ten actors. Both whole face stimuli and half face stimuli are trained upon. Training was stopped when network performance on the training set most closely correlated with the human confusion matrix reported by Ekman.

Experiment 1: Half-Faces Happy Sad Fearful Angry Surprised Disgusted Our first experiment explored recognition of half-face images. An expression should be recognizable in both halves of a composite image in order for one half to interfere with the other half. How well does our model correspond to Calder’s results using humans? The half-face test stimuli are created by zeroing the Gabor outputs in one half (top/bottom) of the image (in the “attenuate” part of the preprocessing). Attenuate Centered on bottom half Centered on bottom half

Experiment 1 Results Fraction of Stimuli Incorrectly Identified (Ten trials of ten networks, standard deviation in parentheses) Expression Human Top Network Top Human Bottom Network Bottom Happy 0.20 (.09) 0.40 0.01 (.01) 0.00 Sadness 0.19 (.05) 0.28 0.34 (.08) Fear 0.33 (.08) 0.56 (.09) 0.70 Anger 0.28 (.06) 0.29 0.49 (.09) 0.65 Surprise 0.06 (.21) 0.33 (.07) 0.21 Disgust 0.62 (.10) 0.20 0.04 (.14)

Experiment 2: Does Expression Use a Configural Model? In this experiment, we test the networks on stimuli created from composite and noncomposite images. Composite/noncomposite images are created using the top halves of images of top-biased expressions and the bottom halves of images of bottom-biased expressions. To focus the network’s attention on one half of the face, the other half is attenuated. attenuate Composite Image Gabor Response Vector Composite Stimulus Noncomposite Stimulus shift convolve

Experiment 2 Results Human Results Network Results Composite Noncomposite Composite Noncomposite Human Reaction Time (ms) Fraction incorrectly identified (bars indicate one standard deviation) The network more correctly identified the expression in one half of the noncomposite than the composite images. When identifying the expression in one half of the stimuli, there is incorrect configural information from the other half of composite stimuli, but not the other half of the noncomposite stimuli. If incorrect configural information disrupts expression recognition, then configural information is used by the network, thus the networks use a configural model.

Experiment 3: Can Incorrect Configural Information Disrupt Identity & Expression Recognition Independently? In this experiment, we test the networks on stimuli created from SID,DE, DID,DE, and DID,SE composites. Test stimuli are created as in Experiment 2. Network outputs both identity and expression classifications. Results for all three types of stimuli are compared. Same Identity, Different Expression Different Identity, Same Expression Different Identity, Different Expression

Experiment 3 Results Human Results Network Results Expression Decision Expression Decision Reaction Time (ms) 1 – Average Correct Output Identity Decision Identity Decision SID/DE DID/DE DID/SE SID/DE DID/DE DID/SE Human Reaction Time (ms) Approximation of Reaction Time: 1 – Average Correct Output It is no harder for humans/model to recognize expression in DID,DE than SID,DE. It is no harder for humans/model to recognize identity in DID,DE than DID,SE. Incorrect identity configural information does not disrupt humans/model’s expression recognition, and vice-versa.

Only One Representation? PCA Representation Our model uses one representation for identity and expression. Our results suggest that identity and expression are encoded by different, orthogonal principal components. It is probable that identity and expression representations evolve together to ensure orthogonality.

Conclusions Our model and humans use similar information in expression and identity recognition: Experiment 1 showed that our model found the same expressions top- and bottom-biased as humans. Experiment 2 showed that our model uses configural information for expression recognition. Experiment 3 showed that incorrect configural information can disrupt identity and expression recognition independently. However, as our model uses only one representation for both identity and expression recognition, it is possible that the same representation be used for both to obtain the same results as Calder’s experiments.