Modeling Facial Shape and Appearance

Slides:



Advertisements
Similar presentations
The Extended Cohn-Kanade Dataset(CK+):A complete dataset for action unit and emotion-specified expression Author:Patrick Lucey, Jeffrey F. Cohn, Takeo.
Advertisements

Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
FACIAL EMOTION RECOGNITION BY ADAPTIVE PROCESSING OF TREE STRUCTURES Jia-Jun Wong and Siu-Yeung Cho Forensic and Security Lab School of Computer Engineering.
DONG XU, MEMBER, IEEE, AND SHIH-FU CHANG, FELLOW, IEEE Video Event Recognition Using Kernel Methods with Multilevel Temporal Alignment.
Face Recognition. Introduction Why we are interested in face recognition? Why we are interested in face recognition? Passport control at terminals in.
Matthias Wimmer, Ursula Zucker and Bernd Radig Chair for Image Understanding Computer Science Technische Universität München { wimmerm, zucker, radig
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
Amir Hosein Omidvarnia Spring 2007 Principles of 3D Face Recognition.
Adviser:Ming-Yuan Shieh Student:shun-te chuang SN:M
Tracking Multiple Occluding People by Localizing on Multiple Scene Planes Professor :王聖智 教授 Student :周節.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
Automatic Analysis of Facial Expressions: The State of the Art Automatic Analysis of Facial Expressions: The State of the Art By Maja Pantic, Leon Rothkrantz.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Meta-Cognition, Motivation, and Affect PSY504 Spring term, 2011 March 16, 2010.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Recent Developments in Human Motion Analysis
Rodent Behavior Analysis Tom Henderson Vision Based Behavior Analysis Universitaet Karlsruhe (TH) 12 November /9.
TAUCHI – Tampere Unit for Computer-Human Interaction Automated recognition of facial expressi ns and identity 2003 UCIT Progress Report Ioulia Guizatdinova.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
1 Modeling Facial Shape and Appearance M. L. Gavrilova.
Human Emotion Synthesis David Oziem, Lisa Gralewski, Neill Campbell, Colin Dalton, David Gibson, Barry Thomas University of Bristol, Motion Ripper, 3CR.
Irfan Essa, Alex Pentland Facial Expression Recognition using a Dynamic Model and Motion Energy (a review by Paul Fitzpatrick for 6.892)
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Graphite 2004 Statistical Synthesis of Facial Expressions for the Portrayal of Emotion Lisa Gralewski Bristol University United Kingdom
A Method for Hand Gesture Recognition Jaya Shukla Department of Computer Science Shiv Nadar University Gautam Budh Nagar, India Ashutosh Dwivedi.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
Multimodal Information Analysis for Emotion Recognition
Object Recognition in Images Slides originally created by Bernd Heisele.
Learning to perceive how hand-written digits were drawn Geoffrey Hinton Canadian Institute for Advanced Research and University of Toronto.
1 Ying-li Tian, Member, IEEE, Takeo Kanade, Fellow, IEEE, and Jeffrey F. Cohn, Member, IEEE Presenter: I-Chung Hung Advisor: Dr. Yen-Ting Chen Date:
Computer Vision Michael Isard and Dimitris Metaxas.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Automatic Facial Emotion Recognition Aitor Azcarate Felix Hageloh Koen van de Sande Roberto Valenti Supervisor: Nicu Sebe.
EMPATH: A Neural Network that Categorizes Facial Expressions Matthew N. Dailey and Garrison W. Cottrell University of California, San Diego Curtis Padgett.
Face Image-Based Gender Recognition Using Complex-Valued Neural Network Instructor :Dr. Dong-Chul Kim Indrani Gorripati.
Face Recognition Summary –Single pose –Multiple pose –Principal components analysis –Model-based recognition –Neural Networks.
Face Detection 蔡宇軒.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
CIRP Annals - Manufacturing Technology 60 (2011) 1–4 Augmented assembly technologies based on 3D bare-hand interaction S.K. Ong (2)*, Z.B. Wang Mechanical.
Data Mining for Surveillance Applications Suspicious Event Detection
Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images
Do Expression and Identity Need Separate Representations?
Computer vision: models, learning and inference
Visual Information Retrieval
Guillaume-Alexandre Bilodeau
Recognizing Deformable Shapes
Tracking Objects with Dynamics
Proposed architecture of a Fully Integrated
Advanced Computer Animation Techniques
EMOTIONAL INTELLIGENCE
Video-based human motion recognition using 3D mocap data
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Multiscale Feature Identification in the Solar Atmosphere
The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression By: Patrick Lucey, Jeffrey F. Cohn, Takeo.
In summary C1={skin} C2={~skin} Given x=[R,G,B], is it skin or ~skin?
Tremor Detection Using Motion Filtering and SVM Bilge Soran, Jenq-Neng Hwang, Linda Shapiro, ICPR, /16/2018.
Data Mining for Surveillance Applications Suspicious Event Detection
Introduction Computer vision is the analysis of digital images
Domingo Mery Department of Computer Science
Rachael E. Jack, Oliver G.B. Garrod, Philippe G. Schyns 
Multimodal Caricatural Mirror
CS4670: Intro to Computer Vision
Data Mining for Surveillance Applications Suspicious Event Detection
Introduction to Object Tracking
Introduction Computer vision is the analysis of digital images
Recognizing Deformable Shapes
Domingo Mery Department of Computer Science
Presentation transcript:

Modeling Facial Shape and Appearance M. L. Gavrilova

Outline Facial Expression Analysis Principles of Facial Expression Analysis Problem Space for Facial Expression Analysis Recent Advances in Automatic Facial Expression Analysis Conclusions

Facial Expression Analysis Facial expressions are the facial changes in response to a person’s internal emotional states, intentions or social communications. Facial expression analysis has been an active research topic for behavioral scientists since the work of Darwin in 1872. Suwa et al. presented an early attempt to automatically analyze facial expressions by tracking the motion of 20 identified spots on an image sequence in 1978.

Principles of Facial Expression Analysis Facial expression analysis refers to computer systems that attempt to automatically analyze and recognize facial motions and facial feature changes from visual information. Sometimes the facial expression analysis has been confused with emotion analysis in the computer vision domain. For emotion analysis, higher level knowledge is required. For example, although facial expressions can convey emotion, they can also express intention, cognitive processes, physical effort, or other intra- or interpersonal meanings.

Principles of Facial Expression Analysis The accomplishments in the related areas such as psychological studies, human movement analysis, face detection, face tracking and recognition make the automatic facial expression analysis possible. Automatic facial expression analysis can be applied in many areas such as: emotion and paralinguistic communication, clinical psychology, psychiatry, neurology, pain assessment, lie detection, intelligent environments, and multimodal human-computer interface (HCI).

Basic Structure of Facial Expression Analysis Systems The general approach to automatic facial expression analysis (AFEA) consists of three steps: face acquisition, facial data extraction and representation, and facial expression recognition. Basic structure of facial expression analysis systems.

Basic Structure of Facial Expression Analysis Systems After the face is located, the next step is to extract and represent the facial changes caused by facial expressions. In facial feature extraction for expression analysis, there are mainly two approaches: geometric feature-based methods present the shape and locations of facial components (including the mouth, eyes, brows, and nose). The facial components or facial feature points are extracted to form a feature vector that represents the face geometry. appearance- based methods apply image filters such as Gabor wavelets to either the whole face or specific regions in a face image to extract a feature vector

Facial Expression Analysis Systems Facial expression recognition is the last stage of the AFEA systems. The facial changes can be identified as facial action units or prototypic emotional expressions. Depending on if the temporal information is used, the recognition approaches can be classified as frame-based or sequence-based. First we discuss the general structure of AFEA systems. Next we describe the problem space for facial expression analysis. This space includes multiple dimensions: level of description, individual differences in subjects, transitions among expressions, intensity of facial expression, deliberate versus spontaneous expression, head orientation and scene complexity, image acquisition and resolution, reliability of ground truth, databases, and the relation to other facial behaviors or nonfacial behaviors. The last part is devoted to a description of more specific approaches and the techniques used in recent advances.

Outline Facial Expression Analysis Principles of Facial Expression Analysis Problem Space for Facial Expression Analysis Recent Advances in Automatic Facial Expression Analysis Conclusions

Problem Space for Facial Expression Analysis With few exceptions, most AFEA systems attempt to recognize a small set of prototypic emotional expressions (e.g. disgust, fear, joy, surprise, sadness, anger). This practice may follow from the work of Drawin, and more recently Ekman and Friesen and Izard et al., who proposed that emotion-specified expressions have corresponding prototypic facial expressions. In everyday life, however, such prototypic expressions occur relatively infrequently. Instead, emotion more often is communicated by subtle changes in one or a few discrete facial features, such as tightening of the lips in anger or obliquely lowering the lip corners in sadness. Emotion-specified facial expression. 1. disgust 2. fear 3. joy 4. surprise 5. sadness 6. anger

Level of Description To capture such subtlety of human emotion and paralinguistic communication, automated recognition of fine-grained changes in facial expression is needed. The facial action coding system (FACS) is a human-observer-based system designed to detect subtle changes in facial features. Viewing videotaped facial behavior in slow motion, trained observers can manually code all possible facial displays, which are referred to as action units and may occur individually or in combinations. FACS consists of 44 action units. Thirty are anatomically related to contraction of a specific set of facial muscles.

Level of Description FACS itself is purely descriptive and includes no inferential labels. By converting FACS codes to EMFACS or similar systems, face images may be coded for emotion-specified expressions as well as for more molar categories of positive or negative emotion.

Properties of an Ideal Facial Expression Analysis System

Outline Facial Expression Analysis Principles of Facial Expression Analysis Problem Space for Facial Expression Analysis Recent Advances in Automatic Facial Expression Analysis Conclusions

Recent Advances in Automatic Facial Expression Analysis The recent research in automatic facial expression analysis tends to follow these directions: Build more robust systems for face acquisition, facial data extraction and representation, and facial expression recognition to handle head motion (in-plane and out-of-plane), occlusion, lighting changes, and lower intensity of expressions. Employ more facial features to recognize more expressions and to achieve a higher recognition rate. Recognize facial action units and their combinations rather than emotion-specified expressions. Recognize action units as they occur spontaneously. Develop fully automatic and real-time AFEA systems.

Recent Advances in Automatic Facial Expression Analysis 2D Image-based Method: To handle the full range of head motion for expression analysis, Tian et al. detected the head instead of the face. The head detection uses the smoothed silhouette of the foreground object as segmented using background subtraction and computing the negative curvature minima (NCM) points of the silhouette.

Recent Advances in Automatic Facial Expression Analysis AEFA Facial Feature Extraction and Representation Two types of features can be extracted: geometric features and appearance features. Geometric features present the shape and locations of facial components (including mouth, eyes, brows, nose). The facial components or facial feature points are extracted to form a feature vector that represents the face geometry. To recognize facial expressions, an AFEA system can use geometric features only, appearance features only or hybrid features (both geometric and appearance features). Research shows that using hybrid features can achieve better results for some expressions. To remove the effects of variation in face scale, motion, lighting and other factors, one can first align and normalize the face to a standard face (2D or 3D) manually or automatically and then obtain normalized feature measurements by using a reference image (neutral face).

Geometric Feature Extraction A three-state lip model describes the lip state: open, closed, tightly closed. A two-state model (open or closed) is used for each of the eyes. Each brow and cheek has a one-state model. Some appearance features, such as nasolabial furrows and crows-feet wrinkles are represented explicitly by using two states: present and absent. Given an image sequence, the region of the face and approximate location of individual face features are detected automatically in the initial frame. The contours of the face features and components then are adjusted manually in the initial frame. Feature extraction of UIUC S1. a. input video frame. b. Snapshot of the geometric tracking system. c. Extracted texture map. d. Selected facial regions for appearance feature extraction.

Appearance Feature Extraction Gabor wavelets are widely used to extract the facial appearance changes as a set of multiscale and multiorientation coefficients. The Gabor filter may be applied to specific locations on a face or to the whole face image. The recognition rates for six emotion- specified expressions (e.g. joy and anger) were significantly higher for Gabor wavelet coefficients. Donato et al. compared several techniques for recognizing six single upper face AUs and six lower face AUs. The best performances were obtained using a Gabor wavelet representation and independent component analysis. Gabor appearance feature extraction in the UCSD systems.

Recent Advances in Automatic Facial Expression Analysis Facial Expression Recognition The last step of AFEA systems is to recognize facial expression based on the extracted features. Many classifiers have been applied to expression recognition such as K-nearest neighbor, multinomial logistic ridge regression (MLR), hidden Markov models (HMM), tree augmented naïve Bayes and others. The frame-based recognition method uses only the current frame with or without a reference image (mainly it is a neutral face image) to recognize the expressions of the frame. The sequence-based recognition method uses the temporal information of the sequences to recognize the expressions for one or more frames. The Table summarizes the recognition methods, recognition rates, recognition outputs and the databases used in the most recent systems.

Summary of Recent Advances FACS AU or expression recognition of recent advances. SVM, support vector machines; MLR, multinomial logistic ridge regression; HMM, hidden Markov models; BN, Bayesian network; GMM, Gaussian mixture model.

Neural network-based recognizer Neural network-based recognizer for AU combinations in CMU S1.

Conclusions for Facial Expression Analysis Four recent trends in automatic facial expression analysis are: diversity of facial features in an effort to increase the number of expressions that may be recognized. recognition of facial action units and their combinations rather than more global and easily identified emotion-specified expressions. more robust systems for face acquisition, facial data extraction and representation, and facial expression recognition to handle head motion (both in-plane and out-of-plane), occlusion, lighting change and low intensity expressions, all of which are common in spontaneous facial behavior in naturalistic environments. fully automatic and real-time AFEA systems.

Conclusions for Facial Expression Analysis Although many recent advances and successes in automatic facial expression analysis have been achieved, many questions remain open. Some major ones are:- How do humans correctly recognize facial expressions? Is it always better to analyze finer levels of expression? Is there any better way to code facial expressions for computer systems? How do we obtain reliable ground truth? How do we recognize facial expressions in real life? Ho do we best use the temporal information? How may we integrate facial expression analysis with other modalities?