Helsinki University of Technology Laboratory of Computational Engineering Modeling facial expressions for Finnish talking head Michael Frydrych, LCE, 11.6.2004.

Slides:



Advertisements
Similar presentations
National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
A. Hatzis, P.D. Green, S. Howard (1) Optical Logo-Therapy (OLT) : Visual displays in practical auditory phonetics teaching. Introduction What.
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Created By: Lauren Snyder, Juliana Gerard, Dom Williams, and Ryan Holsopple.
Digital Interactive Entertainment Dr. Yangsheng Wang Professor of Institute of Automation Chinese Academy of Sciences
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
© Copyright IBM Corporation 2006 Cell Broadband Engine Applications Francesco Bertagnolli System & Technology Group Examples of use.
 Public speaking is speaking to a group of people in a structured, deliberate manner intended to inform, influence, or entertain the listeners.
Facial expression as an input annotation modality for affective speech-to-speech translation Éva Székely, Zeeshan Ahmed, Ingmar Steiner, Julie Carson-Berndsen.
Speech Group INRIA Lorraine
3D Face Modeling Michaël De Smet.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
Dmitri Bitouk Shree K. Nayar Columbia University Creating a Speech Enabled Avatar from a Single Photograph.
LYU0603 A Generic Real-Time Facial Expression Modelling System Supervisor: Prof. Michael R. Lyu Group Member: Cheung Ka Shun ( ) Wong Chi Kin ( )
Human Face Modeling and Animation Example of problems in application of multimedia signal processing.
MUltimo3-D: a Testbed for Multimodel 3-D PC Presenter: Yi Shi & Saul Rodriguez March 14, 2008.
High-Quality Video View Interpolation
Computer Science Department Graduate Orientation 1 August 31, 2006 Center for Visual Computing.
Artificial Intelligence & Information Analysis Group (AIIA) Centre of Research and Technology Hellas INFORMATICS & TELEMATICS INSTITUTE.
CSCE 641: Computer Graphics Image-based Rendering Jinxiang Chai.
CSD 5230 Advanced Applications in Communication Modalities 7/3/2015 AAC 1 Introduction to AAC Orientation to Course Assessment Report Writing.
Convergence of vision and graphics Jitendra Malik University of California at Berkeley Jitendra Malik University of California at Berkeley.
BASIC CONCEPS OF ANIMATION The presentation source: Department of Educational Multimedia Faculty of Education, UTM MPT 1383: VIDEO AND ANIMATION TECHNOLOGY.
Computer Animation Prepared by Khadija Kuhail Supervised by Dr Sanaa Alsayegh.
22C:196 Computer Animation Instructor: Joe Kearney 101M MLH
Database Construction for Speech to Lip-readable Animation Conversion Gyorgy Takacs, Attila Tihanyi, Tamas Bardi, Gergo Feldhoffer, Balint Srancsik Peter.
Perception-Based Engineering: Integrating Human Response Into Product Design A collaboration between psychologists and engineers. Mission: To integrate.
Assistive Technology and Education Mrs. G. Bacal Guidelines Designed for people who struggle to learn for different reasons, such as: learning disabilities,emotional.
Laboratory of Computational Engineering Michael Frydrych, Making the head smile Smile? -> part of non-verbal communication Functions of non-verbal.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Computer Graphics Computer Graphics is everywhere: Visual system is most important sense: High bandwidth Natural communication Fast developments in Hardware.
Use and Re-use of Facial Motion Capture M. Sanchez, J. Edge, S. King and S. Maddock.
Facial animation retargeting framework using radial basis functions Tamás Umenhoffer, Balázs Tóth Introduction Realistic facial animation16 is a challenging.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
Real-Time Animation of Realistic Virtual Humans. 1. The 3D virtual player is controlled by the real people who has a HMD and many sensors people who has.
APML, a Markup Language for Believable Behavior Generation Soft computing Laboratory Yonsei University October 25, 2004.
Computer Graphics 2 In the name of God. Outline Introduction Animation The most important senior groups Animation techniques Summary Walking, running,…examples.
110/20/ :06 Graphics II Paper Reviews Facial Animation Session 8.
N n Debanga Raj Neog, Anurag Ranjan, João L. Cardoso, Dinesh K. Pai Sensorimotor Systems Lab, Department of Computer Science The University of British.
Speech Perception 4/4/00.
Computer Science Department Pacific University Artificial Intelligence -- Computer Vision.
Lecture 15 – Social ‘Robots’. Lecture outline This week Selecting interfaces for robots. Personal robotics Chatbots AIML.
1 Reconstructing head models from photograph for individualized 3D-audio processing Matteo Dellepiane, Nico Pietroni, Nicolas Tsingos, Manuel Asselot,
Intelligent Robot Architecture (1-3)  Background of research  Research objectives  By recognizing and analyzing user’s utterances and actions, an intelligent.
Perceptual Analysis of Talking Avatar Head Movements: A Quantitative Perspective Xiaohan Ma, Binh H. Le, and Zhigang Deng Department of Computer Science.
4 November 2000Bridging the Gap Workshop 1 Control of avatar gestures Francesca Barrientos Computer Science Division UC Berkeley.
Animated Speech Therapist for Individuals with Parkinson Disease Supported by the Coleman Institute for Cognitive Disabilities J. Yan, L. Ramig and R.
S PEECH T ECHNOLOGY Answers to some Questions. S PEECH T ECHNOLOGY WHAT IS SPEECH TECHNOLOGY ABOUT ?? SPEECH TECHNOLOGY IS ABOUT PROCESSING HUMAN SPEECH.
MMDB-9 J. Teuhola Standardization: MPEG-7 “Multimedia Content Description Interface” Standard for describing multimedia content (metadata).
1 Galatea: Open-Source Software for Developing Anthropomorphic Spoken Dialog Agents S. Kawamoto, et al. October 27, 2004.
Performance Driven Facial Animation
Electronic visualization laboratory, university of illinois at chicago Towards Lifelike Interfaces That Learn Jason Leigh, Andrew Johnson, Luc Renambot,
Facial Motion Cloning Using Global Shape Deformation Marco Fratarcangeli and Marco Schaerf University of Rome “La Sapienza”
Facial Animation Wilson Chang Paul Salmon April 9, 1999 Computer Animation University of Wisconsin-Madison.
Observing Lip and Vertical Larynx Movements During Smiled Speech (and Laughter) - work in progress - Sascha Fagel 1, Jürgen Trouvain 2, Eva Lasarcyk 2.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
Animation Animation is about bringing things to life Technically: –Generate a sequence of images that, when played one after the other, make things move.
UCL Human Representation in Immersive Space. UCL Human Representation in Immersive Space Body ChatSensing Z X Y Zr YrXr Real–Time Animation.
“Articulatory Talking Head” Showcase Project, INRIA, KTH. Articulatory Talking Head driven by Automatic Speech Recognition INRIA, Parole Team KTH, Centre.
Computer Animation Algorithms and Techniques
Visual Information Retrieval
Chapter 15 Gestures and Sign Languages
MikeTalk:An Adaptive Man-Machine Interface
Assistive System Progress Report 1
University of West Bohemia – Department of Cybernetics
AN INTRODUCTION TO COMPUTER GRAPHICS Subject: Computer Graphics Lecture No: 01 Batch: 16BS(Information Technology)
Computer Graphics Lecture 15.
Lecture 3. Virtual Worlds : Representation,Creation and Simulation ( II ) 고려대학교 그래픽스 연구실.
Presentation transcript:

Helsinki University of Technology Laboratory of Computational Engineering Modeling facial expressions for Finnish talking head Michael Frydrych, LCE,

Laboratory of Computational Engineering Michael Frydrych, Finnish talking head Computer animated model of a talking person Synchronized A/V speech Model of emotional facial expressions

Laboratory of Computational Engineering Michael Frydrych, User interface of “old” talking head

Laboratory of Computational Engineering Michael Frydrych, What has been done with it? Studies in audiovisual speech perception Kiosk-interface at the University of Tampere Cultural activities Major role in play Kyberias at Kellariteatteri (2001) Talking Head

Laboratory of Computational Engineering Michael Frydrych, Content Talking heads – why? Animation methods Controlling animation Making them speak Practicals Making the head smile Emotions –why? Practicals

Laboratory of Computational Engineering Michael Frydrych, Why talking heads? Entertainment Information services Ananova, information kiosks Education services Learning foreign languages,… Agents in spoken dialogue systems nonverbal signals, comfort

Laboratory of Computational Engineering Michael Frydrych, Tampere museums Tampere museums

Laboratory of Computational Engineering Michael Frydrych, Aids in communication Speech is both heard and seen Improve intelligibility in noisy env. Aid for hearing impaired people Synface

Laboratory of Computational Engineering Michael Frydrych, Synface (telephone -> animated face) Figure by KTH Stockholm

Laboratory of Computational Engineering Michael Frydrych, … applications Language training speech training for profoundly deaf Diagnostics and therapy EU: VEPSY, VREPAR (assess and treat naxiety disorders and specific phobia)

Laboratory of Computational Engineering Michael Frydrych, Audiovisual speech integration = combining auditory and visual percepts into a single speech percept Strength of integration is demonstrated by McGurk- effect: combining sound /pa/ to a face ”telling” /ka/, speech percept is often /ta/ (McGurk & MacDonald, 1976, Nature) /pa/ /ka/ /ta/

Laboratory of Computational Engineering Michael Frydrych, Result:Computer animated talking face improves intelligibility of auditory speech A study in audio-visual speech perception

Laboratory of Computational Engineering Michael Frydrych, Psychophysical and psychophysiological experiments Audiovisual speech perception Emotion research … Benefits Natural stimuli may contain unwanted features Full controllability Quick creation of stimuli … application in research

Laboratory of Computational Engineering Michael Frydrych, Bulding on realism Realism: 1) Objective  topography, animation, texture, synchronization,... 2) Subjective (communication)  Audio-visual speech  Facial expressions, nonverbal behavior (prosody, eye movements) Evaluation: Objective  Subjective

Laboratory of Computational Engineering Michael Frydrych, Making the head speak Issues: Voice - speech synthesizer Animation – parameterization Synchronization

Laboratory of Computational Engineering Michael Frydrych, Acoustic Speech Generation Based on Festival platform. Developed at The Centre for Speech Technology Research, University of Edinburg, England. Scheme programming language, allows to program behaviour Finnish voice, prosody, expansion (numerals, etc.) Department of Phonetics, University of Helsinki Issues: production of articulatory parameters, synchronization

Laboratory of Computational Engineering Michael Frydrych, Animation methods - representation Polygonal Keyframing libraries of postures, interpolation Parametric deformations deformations are grouped under parameters meaningful to the animator Muscle-Based deformations Interactive deformations numerous control points, deformation propagation Free Form deformations deformation associated with a deformation box

Laboratory of Computational Engineering Michael Frydrych, Splines Implicit surfaces Physics-based models Physical models of the skin Volume preservation Deformations by inducing forces

Laboratory of Computational Engineering Michael Frydrych, Hooks to data Need the geometry of faces Rendering properties Deformation of facial expression or speech How? 2D and 3D techniques

Laboratory of Computational Engineering Michael Frydrych, D Input 3D digitizer is the most direct way, fairly automatic (Optotrack) 3D trackers – digitizing of projected/marked mesh, rather manual CT (Computer Tomography) and MRI (Magnetic Resonance Imaging) and … 3D modeling programs

Laboratory of Computational Engineering Michael Frydrych, D Input Photogrammetry Two images of an object are taken from different viewpoints, corresponding points are found The 3D shape of faces can be determined from a single 2D image after projecting of regular pattern Generic facial model is prepared and transformed to “match” a photograph 3rd dimension can be approximated by acquiring face model (set priors) and Bayesian inference

Laboratory of Computational Engineering Michael Frydrych, Texture mapping

Laboratory of Computational Engineering Michael Frydrych, break

Laboratory of Computational Engineering Michael Frydrych, Data for articulation and expressions Keyframing -> expression libraries Real-time/performance data Parameterization Articulatory parameters – jaw opening, lip rounding, lip protrusion, … Facial expressions – FACS Statistical models from expression libraries or real-time data

Laboratory of Computational Engineering Michael Frydrych, Statistical parameterization Parameterized model learned from 3D performance data (Reveret) Figure by ISCP Grenoble

Laboratory of Computational Engineering Michael Frydrych, … three control parameters Figure by ISCP Grenoble

Laboratory of Computational Engineering Michael Frydrych, … and the results Jaw Rounding Opening Raising Figure by ISCP Grenoble

Laboratory of Computational Engineering Michael Frydrych, Video by ISCP Grenoble

Laboratory of Computational Engineering Michael Frydrych, Finnish talking head Audiovisual database Using MaxReflex 3D optical tracker (at Linköping Univ.) Multiple IR cameras, reflexive markers  reconstruction from stereo Coarticulation, lips, visual prosody

Laboratory of Computational Engineering Michael Frydrych, Point-lights positions

Laboratory of Computational Engineering Michael Frydrych, Demo – live recording at Linköping

Laboratory of Computational Engineering Michael Frydrych, How to create “visemes” ?

Laboratory of Computational Engineering Michael Frydrych, Demo – reconstructed motion 10 fps 40 fps

Laboratory of Computational Engineering Michael Frydrych, Figure by ISCP Grenoble

Laboratory of Computational Engineering Michael Frydrych,

Laboratory of Computational Engineering Michael Frydrych,

Laboratory of Computational Engineering Michael Frydrych, End of 1st part