Perceptually Guided Expressive Facial Animation Zhigang Deng and Xiaohan Ma Computer Graphics and Interactive Media Lab Department of Computer Science.

Slides:



Advertisements
Similar presentations
Towards Human-Realistic Animation by Observing Real Human Dynamics Darren Cosker Towards Human-Realistic Animation by Observing Real Human Dynamics Darren.
Advertisements

Affective Facial Expressions Facilitate Robot Learning Joost Broekens Pascal Haazebroek LIACS, Leiden University, The Netherlands.
The Chinese Room: Understanding and Correcting Machine Translation This work has been supported by NSF Grants IIS Solution: The Chinese Room Conclusions.
Rui Tang, Darren Cosker and Wenbin Li Global Alignment for Dynamic 3D Morphable Model Construction University of Bath.
The Extended Cohn-Kanade Dataset(CK+):A complete dataset for action unit and emotion-specified expression Author:Patrick Lucey, Jeffrey F. Cohn, Takeo.
Detail to attention: Exploiting Visual Tasks for Selective Rendering Kirsten Cater 1, Alan Chalmers 1 and Greg Ward 2 1 University of Bristol, UK 2 Anyhere.
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
Face Alignment with Part-Based Modeling
Facial expression as an input annotation modality for affective speech-to-speech translation Éva Székely, Zeeshan Ahmed, Ingmar Steiner, Julie Carson-Berndsen.
Adviser:Ming-Yuan Shieh Student:shun-te chuang SN:M
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
1 Face Synthesis M. L. Gavrilova. 2 Outline Face Synthesis From Modeling to Synthesis Facial Expression Synthesis Conclusions.
Retargeting Algorithms for Performance-Driven Animation J.P. Lewis Fred Pighin.
Biometrics: Faces and Identity Verification in a Networked World
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
The Use of Eye Tracking Technology in the Evaluation of e-Learning: A Feasibility Study Dr Peter Eachus University of Salford.
Principal Component Analysis
LYU0603 A Generic Real-Time Facial Expression Modelling System Supervisor: Prof. Michael R. Lyu Group Member: Cheung Ka Shun ( ) Wong Chi Kin ( )
Recent Developments in Human Motion Analysis
RECOGNIZING FACIAL EXPRESSIONS THROUGH TRACKING Salih Burak Gokturk.
Animation From Motion Capture Motion Capture Assisted Animation: Texturing and Synthesis Kathy Pullen Chris Bregler Motion Capture Assisted Animation:
A Data-Driven Approach to Quantifying Natural Human Motion SIGGRAPH ’ 05 Liu Ren, Alton Patrick, Alexei A. Efros, Jassica K. Hodgins, and James M. Rehg.
Automatic Pose Estimation of 3D Facial Models Yi Sun and Lijun Yin Department of Computer Science State University of New York at Binghamton Binghamton,
Feature Detection and Emotion Recognition Chris Matthews Advisor: Prof. Cotter.
Constraint-based Motion Optimization Using A Statistical Dynamic Model Jinxiang Chai Texas A&M University.
1 Expression Cloning Jung-yong Noh Ulrich Neumann Siggraph01.
Dynamic Response for Motion Capture Animation Victor B. Zordan Anna Majkowska Bill Chiu Matthew Fast Riverside Graphics Lab University of California, Riverside.
Presented by Zeehasham Rasheed
Vision-based Control of 3D Facial Animation Jin-xiang Chai Jing Xiao Jessica Hodgins Carnegie Mellon University.
Artificial Intelligence & Information Analysis Group (AIIA) Centre of Research and Technology Hellas INFORMATICS & TELEMATICS INSTITUTE.
Facial Type, Expression, and Viseme Generation Josh McCoy, James Skorupski, and Jerry Yee.
Facial Type, Expression, and Viseme Generation Josh McCoy, James Skorupski, and Jerry Yee.
Interactive Optimization by Genetic Algorithms Cases: Lighting Patterns and Image Enhancement Janne Koljonen Electrical Engineering and Automation, University.
Sunee Holland University of South Australia School of Computer and Information Science Supervisor: Dr G Stewart Von Itzstein.
Computer Vision Systems for the Blind and Visually Disabled. STATS 19 SEM Talk 3. Alan Yuille. UCLA. Dept. Statistics and Psychology.
Toshiba Update 04/09/2006 Data-Driven Prosody and Voice Quality Generation for Emotional Speech Zeynep Inanoglu & Steve Young Machine Intelligence Lab.
Performance Evaluation of Grouping Algorithms Vida Movahedi Elder Lab - Centre for Vision Research York University Spring 2009.
Human Emotion Synthesis David Oziem, Lisa Gralewski, Neill Campbell, Colin Dalton, David Gibson, Barry Thomas University of Bristol, Motion Ripper, 3CR.
EWatchdog: An Electronic Watchdog for Unobtrusive Emotion Detection based on Usage Analysis Rayhan Shikder Department.
Efficient Editing of Aged Object Textures By: Olivier Clément Jocelyn Benoit Eric Paquette Multimedia Lab.
Use and Re-use of Facial Motion Capture M. Sanchez, J. Edge, S. King and S. Maddock.
Facial animation retargeting framework using radial basis functions Tamás Umenhoffer, Balázs Tóth Introduction Realistic facial animation16 is a challenging.
Graphite 2004 Statistical Synthesis of Facial Expressions for the Portrayal of Emotion Lisa Gralewski Bristol University United Kingdom
Plausible motion: conclusion Ronen Barzel John Hughes.
Vehicle License Plate Detection Algorithm Based on Statistical Characteristics in HSI Color Model Instructor : Dr. K. R. Rao Presented by: Prasanna Venkatesh.
N n Debanga Raj Neog, Anurag Ranjan, João L. Cardoso, Dinesh K. Pai Sensorimotor Systems Lab, Department of Computer Science The University of British.
Mingyang Zhu, Huaijiang Sun, Zhigang Deng Quaternion Space Sparse Decomposition for Motion Compression and Retrieval SCA 2012.
Spatio-Temporal Free-Form Registration of Cardiac MR Image Sequences Antonios Perperidis s /02/2006.
Perceptual Analysis of Talking Avatar Head Movements: A Quantitative Perspective Xiaohan Ma, Binh H. Le, and Zhigang Deng Department of Computer Science.
Look Over Here: Attention-Directing Composition of Manga Elements Ying Cao Rynson W.H. Lau Antoni B. Chan SIGGRAPH
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Mining Logs Files for Data-Driven System Management Advisor.
Design of PCA and SVM based face recognition system for intelligent robots Department of Electrical Engineering, Southern Taiwan University, Tainan County,
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 14 November 2003 William.
University of Washington v The Hebrew University * Microsoft Research Synthesizing Realistic Facial Expressions from Photographs Frederic Pighin Jamie.
Unsupervised Learning for Speech Motion Editing Eurographics/SIGGRAPH Symposium on Computer Animation (2003) Yong Cao 1,2 Petros Faloutsos 1 Frederic.
Language in Cognitive Science. Research Areas for Language Computational models of speech production and perception Signal processing for speech analysis,
Facial Motion Cloning Using Global Shape Deformation Marco Fratarcangeli and Marco Schaerf University of Rome “La Sapienza”
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Animation From Observation: Motion Editing Dan Kong CMPS 260 Final Project.
Immersive Virtual Characters for Educating Medical Communication Skills J. Hernendez, A. Stevens, D. S. Lind Department of Surgery (College of Medicine)
Face Detection 蔡宇軒.
V k equals the vector difference between the object and the block across the first and last frames in the image sequence or more formally: Toward Learning.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
Qian Yu 1, 2, Qing Li 1, Zhigang Deng 1 1 University of Houston
Computer Vision, Robotics, Machine Learning and Control Lab
Fast Preprocessing for Robust Face Sketch Synthesis
Enhanced-alignment Measure for Binary Foreground Map Evaluation
Authoring Directed Gaze for Full-Body Motion Capture
CSc4730/6730 Scientific Visualization
Easy Generation of Facial Animation Using Motion Graphs
Presentation transcript:

Perceptually Guided Expressive Facial Animation Zhigang Deng and Xiaohan Ma Computer Graphics and Interactive Media Lab Department of Computer Science University of Houston

Talk Outline Motivation Related Work Our Approach –Construction of facial perceptual metric –Perceptually Guided Facial Animation Algorithms Results and User Studies Conclusions and Discussion

Motivations How to efficiently measure and synthesize realistic expressive facial animation? –The ultimate measuring stick is human perception. –Intrinsic sensitivity to the subtlety of animated faces Current popular means is to conduct subjective user studies (offline, post-production evaluation tool) –Not automated (tedious human involvements) –Inefficient (time-consuming experiment setup and user study) –Costly (participant cost)

Related Work Facial animation techniques –Geometric deformation [ Singh and Fiume 98, Noh and Neumann 01,Sumner and Popovic 04 ], physically-based [ Lee et al 95, Sifakis et al. 05 ], performance driven facial animation [ Williams 90 ], facial expression synthesis and editing [ zhang et al. 03, Joshi et al. 03 ] Data-driven approaches for facial animation [Brand 99, Chuang et al. 02, Vlasic et al. 05,Wampler et al. 07,Bregler et al. 97, Kshirsagar and Thalmann 03, Cao et al. 04, Deng and Neumann 06] –Focus on the mathematical accuracy/efficiency of their algorithms –Little attention has been paid to perceptual aspects to their algorithms Automatic analysis of facial expressions in computer vision community [Pantic and Rothkrantz 00, Tian et al. 01, Valstar and Pantic 06] –Focus on analysis side of facial expressions

Related Work User studies for character animation [Hodgins et al. 98, Sullivan and Dingliana 01, Sullivan et al. 03, Watson et al. 01, Reitsma and Pollard 03, Wang and Bodenheimer 04, McDonnell et al. 06, McDonnell et al. 07] –Measure the association between human perception and factors of character animation Subjective evaluation have been also conducted to gain human perceptual insight on facial animation [Cunningham et al. 03, Cunningham et al. 04, Wallraven et al. 05, Wallraven et al. 08, Geiger et al. 03, Cosker et al. 05] –Most of these efforts are centered at the qualitative side Our work –Aim to quantitatively model (perceptual metric) the association between facial motion and perceptual outcomes –Further exploit this perceptual metric to build perceptually guided expressive facial animation algorithms

Our Work A novel computational facial perceptual metric (FacePEM) –Measuring and predict the expressiveness (type/scale) of synthetic expressive facial animations –Learn statistical perceptual prediction model (FacePEM) which can measure and predict the perceptual outcomes of arbitrary facial motion FacePEM-guided facial animation algorithms –Perceptual metric-guided speech animation synthesis –Expressive facial motion editing enhanced with expressiveness cues

Construction of FacePEM

Data Acquisition Optical motion capture system for facial motion capture 103 facial markers (95 face markers, 4 head markers, and 4 neck markers) Four human subjects spoke sentences with four emotions (happy, angry, sad, and neutral) Remove head motion and align facial motions of different subjects

Construction of FacePEM

User Evaluation Study Participants were required to identify perceived emotion and corresponding emotional expressiveness scale (1 to 10) as a nonforced-choice task after viewing each clip. 68 facial motion clips, 30 participants (university students) Perceptual Outcome Vector (POV) for each facial motion clip

Facial Motion Analysis Face segmentation –PCA is a global transformation; no explicit correspondence between PCA eigen-vectors and localized facial movements. –Physically-motivated segmentation [Joshi et al. 03] : six regions (forehead, eye, left cheek, right cheek, mouth, nose) Region-based motion reduction –Apply PCA for the movements of each region –Region-based PCA eigen-vectors correspond to meaningful, localized facial movements [Li and Deng 07]. The first/second largest eigen-vectors of the mouth region The first/second largest eigen-vectors of the eye region

Facial Motion Modeling Modeling region-based expressive facial motion patterns –Use the M-order Linear Dynamical Systems (LDS) [Pavlovic et al. 00, Chai and Hodgins 07] to model region-based facial motion patterns for any specific emotion –18 LDSs are fitted (6 regions*3 emotions), x n = ∑(A i *x n-i )+v i Objective Matchness Vector (OMV) –Closeness function: Describes the closeness of any given facial motion sequence at a specific facial region representing a specific emotion –A OMV vector encloses 18 components P emo,reg (S i ) = e -E E = -ln F(S) = -ln F(x 1:T ) ≈ C*∑||x t -∑(A j x t-j )-v j || 2

Construction of FacePEM

Learning Perceptual Prediction Statistical perceptual prediction model –Predict its POV, based on the OMV of input facial motion sequence –Three approaches: Least-square based linear fitting, Radial Basis Functions (RBFs) network approach, and Support Vector Machines (SVMs) –SVMs-based perceptual prediction model achieve the minimum error. Cross-Validation for test and validation –54 facial motion clips as the training dataset –14 facial motion clips as the test/validation dataset

Review of FacePEM Framework

Perceptually Guided Facial Animation Algorithms Perceptually guided expressive speech animation synthesis – The core part of many data-driven speech animation synthesis [Brelger et al. 97, Kshirsagar and Thalmann 03, Cao et al. 04, Deng and Neumann 06] –Cost = PhoMtchCost+ConstrCost+SmoCost –More intelligent and smart “cost function” with perceptual metric Expressive facial motion editing enhanced with expressiveness cues –Current facial motion editing techniques [Chuang et al. 02, Cao et al. 03, Joshi et al. 03, Vlasic et al. 05, Li and Deng 07] do not provide feedback or expressiveness cues –FacePEM will measure and display its updated emotion type and expressiveness scale to users in a timely means.

Perceptually Guided Expressive Speech Animation Synthesis Choose an expressive speech animation synthesis algorithm [Deng and Neumann 06] for test/validation Predict emotion type and expressiveness scale of facial motion sequence being synthesized, and Incorporate emotion scale into synthesis algorithm –EC(s,Emo)=C*(1-Same(Emo, EmoLabel(s))) –EC(s, Emo) = C*(1- CalcEmo(s)[Emo])

Facial Motion Editing Enhanced with Expressiveness Cues Choose an expressive facial motion editing system for validation [Li and Deng 07] Predict emotion type and expressiveness scale of the edited facial motion sequence Display expression type/scale to users as feedback

Results & Evaluation - I 20 expressive facial animation clips (10 from this work, 10 from previous work) 20 participants rated the visual fidelity (expression) of these clips

Results & Evaluation - II 10 expressive edited facial motion clips (5 from this work, 5 from previous work) 20 participants rated the visual fidelity (expression) of these clips

Discussion It is hard to know how much data would be enough to train well-behaviored statistical learning approaches We did not consider the effects of eye gaze motion Limited emotion types (3 basic emotion) are studied. Idiosyncratic motion signals might exist in current used facial motion dataset.

Conclusions We present a novel computational perceptual metrics for measuring and predict the expression (type/scale) of facial animations –Bridge human perceptual insights with objective facial motion patterns Demonstrated perceptually guided expressive facial animation algorithms on two cases –Expressive speech animation synthesis –Interactive expressive facial motion editing Future Work –Study more data and expression types –Remove the idiosyncratic components from the data

Questions?