Multimodal Caricatural Mirror

Slides:



Advertisements
Similar presentations
National Technical University of Athens Department of Electrical and Computer Engineering Image, Video and Multimedia Systems Laboratory
Advertisements

1 Multimodal Technology Integration for News-on-Demand SRI International News-on-Demand Compare & Contrast DARPA September 30, 1998.
Österreichisches Forschnungsinstitut für Artificial Intelligence Featuring the GEMEP Corpus Experiences and Future Plans Hannes Pirker OFAI, Vienna.
Matthias Wimmer, Bernd Radig, Michael Beetz Chair for Image Understanding Computer Science TU München, Germany A Person and Context.
Descriptive schemes for facial expression introduction.
Audio-based Emotion Recognition for Advanced Information Retrieval in Judicial Domain ICT4JUSTICE 2008 – Thessaloniki,October 24 G. Arosio, E. Fersini,
Model-based Image Interpretation with Application to Facial Expression Recognition Matthias Wimmer
RRL: A Rich Representation Language for the Description of Agent Behaviour in NECA Paul Piwek, ITRI, Brighton Brigitte Krenn, OFAI, Vienna Marc Schröder,
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
High Level Prosody features: through the construction of a model for emotional speech Loic Kessous Tel Aviv University Speech, Language and Hearing
Vocal Joystick A New Dimension in Human-Machine Interaction ET 2 Presentation Group 3 Jeremy Moody, Carrie Chudy.
DPSS Radar Research Topics 2 December 2010 Leon Staphorst Principle Researcher Radar and Electronic Warfare Defence, Peace, Safety and Security.
Content-based Video Indexing, Classification & Retrieval Presented by HOI, Chu Hong Nov. 27, 2002.
Recent Developments in Human Motion Analysis
Overview of Computer Vision CS491E/791E. What is Computer Vision? Deals with the development of the theoretical and algorithmic basis by which useful.
Vision-based Control of 3D Facial Animation Jin-xiang Chai Jing Xiao Jessica Hodgins Carnegie Mellon University.
Optimal Adaptation for Statistical Classifiers Xiao Li.
Video Mining Learning Patterns of Behaviour via an Intelligent Image Analysis System.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Non-invasive Techniques for Human Fatigue Monitoring Qiang Ji Dept. of Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute
Emotional Intelligence and Agents – Survey and Possible Applications Mirjana Ivanovic, Milos Radovanovic, Zoran Budimac, Dejan Mitrovic, Vladimir Kurbalija,
Track: Speech Technology Kishore Prahallad Assistant Professor, IIIT-Hyderabad 1Winter School, 2010, IIIT-H.
Facial Feature Detection
IIS for Image Processing Michael J. Watts
Enabling enactive interaction in virtualized experiences Stefano Tubaro and Augusto Sarti DEI – Politecnico di Milano, Italy.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
1 Darmstadt, October 02, 2007 Amalia Ortiz Asociación VICOMTech Mikeletegi Pasealekua Donostia - San Sebastián (Gipuzkoa)
Multimodal Interaction Dr. Mike Spann
SIMILAR NoE at the HUMAINE meeting - 5/06/2007 Multimodal Interaction R&D 4 Years Dec 2003 – Dec M€ 32 partners + 8 fellows.
Chapter 7. BEAT: the Behavior Expression Animation Toolkit
Multimodal Information Analysis for Emotion Recognition
Subtask 1.8 WWW Networked Knowledge Bases August 19, 2003 AcademicsAir force Arvind BansalScott Pollock Cheng Chang Lu (away)Hyatt Rick ParentMark (SAIC)
Overview of Part I, CMSC5707 Advanced Topics in Artificial Intelligence KH Wong (6 weeks) Audio signal processing – Signals in time & frequency domains.
 Motivated by desire for natural human-robot interaction  Encapsulates what the robot knows about the human  Identity  Location  Intentions Human.
Cognitive Systems Foresight Language and Speech. Cognitive Systems Foresight Language and Speech How does the human system organise itself, as a neuro-biological.
AUTOMATIC TARGET RECOGNITION AND DATA FUSION March 9 th, 2004 Bala Lakshminarayanan.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
On the relevance of facial expressions for biometric recognition Marcos Faundez-Zanuy, Joan Fabregas Escola Universitària Politècnica de Mataró (Barcelona.
Facial Expression Analysis Theoretical Results –Low-level and mid-level segmentation –High-level feature extraction for expression analysis (FACS – MPEG4.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
Detection Of Anger In Telephone Speech Using Support Vector Machine and Gaussian Mixture Model Prepared By : Siti Marahaini Binti Mahamood.
Emotional Intelligence Vivian Tseng, Matt Palmer, Jonathan Fouk Group #41.
Brief Intro to Machine Learning CS539
Course Outline (6 Weeks) for Professor K.H Wong
Modeling Expressivity in ECAs
Computer Animation Algorithms and Techniques
REAL-TIME DETECTOR FOR UNUSUAL BEHAVIOR
Nicole looks at faces Development of a visual robot interface to interpret facial expressions NICOLE: Future robots will share their environment with humans.
KRISTINA Consortium Presented by: Mónica Domínguez (UPF-TALN)
Derek McColl Alexander Hong Naoaki Hatakeyama Goldie Nejat
National Institute of Standards and Technology (NIST) Advanced Manufacturing Technology Consortia (AMTech) Program Award Number: 70NANB14H056 Development.
SENSOR FUSION LAB RESEARCH ACTIVITIES PART I : DATA FUSION AND DISTRIBUTED SIGNAL PROCESSING IN SENSOR NETWORKS Sensor Fusion Lab, Department of Electrical.
FAULT TOLERANT TECHNIQUES FOR WIRELESS AD HOC SENSOR NETWORKS
AHED Automatic Human Emotion Detection
GESTURE RECOGNITION TECHNOLOGY
IIS for Image Processing
Anne Pratoomtong ECE734, Spring2002
What is Pattern Recognition?
Kocaeli University Introduction to Engineering Applications
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
T H E P U B G P R O J E C T.
Final Project Presentation | CIS3203
Multimodal Caricatural Mirror
Project #2 Multimodal Caricatural Mirror Intermediate report
Presented by: Mónica Domínguez
John H.L. Hansen & Taufiq Al Babba Hasan
EE 492 ENGINEERING PROJECT
THE TOPICS AND TITLES OF RESEARCH
End-to-End Speech-Driven Facial Animation with Temporal GANs
Presentation transcript:

Multimodal Caricatural Mirror Olivier Martin, UCL (Belgium)

Project Goals Create a Multimodal caricatural mirror : Multimodal = facial + vocal Caricatural = Amplify emotions Mirror = Face your avatar! 2/24/2019

Motivations Emotion Recognition  intelligent systems Modelling emotions  emotion synthesis Interactions  real emotions  database Multimodal Gain 2/24/2019

Technical challenges Multimodal Face Tracking Facial features’ extraction Vocal features’ extraction Multimodal emotion recognition Multimodal emotion synthesis 2/24/2019

Multimodal Face Tracking Automatic tracking of the face, based upon: Skin colour information Ellipsoid-shaped properties (Hough transform,…) Luminance/Chrominance gradient Pre-segmentation of user’s body Array of microphones Infering face from facial features… … 2/24/2019

Skin detection 2/24/2019

Trace Transform using luminance gradient 2/24/2019

Technical challenges Multimodal Face Tracking Facial features’ extraction Vocal features’ extraction Multimodal emotion recognition Multimodal emotion synthesis 2/24/2019

Facial features extraction Detect and track facial features : Localization : learning and/or heuristics Extraction : exploiting a priori knowledge Shape/contour information ‘Crucial points’ information (MPEG-4,…) Temporal ripples … 2/24/2019

Facial features’ extraction 2/24/2019

Facial features’ extraction 2/24/2019

‘Emotional Mask’ 2/24/2019

Technical challenges Multimodal Face Tracking Facial features’ extraction Vocal features’ extraction Multimodal emotion recognition Multimodal emotion synthesis 2/24/2019

Vocal features’ extraction Pitch, energy, speaking rate, noise, MFCC,…are related to ‘the way we speak’ (prosody) Statistics about the features (mean, std dev, enveloppe, …) Learning strategy for features’ selection, for each emotion (forward/backward selection) 2/24/2019

Technical challenges Multimodal Face Tracking Facial features’ extraction Vocal features’ extraction Multimodal emotion recognition Multimodal emotion synthesis 2/24/2019

Multimodal emotion recognition Compare monomodal systems’ performances to multimodal system’s performances, for each emotion  build intelligent classifiers How to synchronize the modalities ? Fusion at which level of the decision process ? (signal level vs semantic level) 2/24/2019

Technical challenges Multimodal Face Tracking Facial features’ extraction Vocal features’ extraction Multimodal emotion recognition Multimodal emotion synthesis 2/24/2019

Multimodal emotion synthesis How to amplify the expression of an emotion ? Build an effective and realistic mapping Synchronisation (lips!) 2/24/2019

Real-time aspects Ideally, facial modality should be real-time Ideally, vocal modality should not be real-time  Goal : Minimize the delay between end of user’s actions and system reactions. 2/24/2019

Technology This has to be discussed within the team… Two types of Machine Learning techniques seem efficient & we have skills! : Support Vector Machines Dynamic Bayesian Networks Powerful animation engines (Maya, 3DSMax,…) Communication between modules : OpenInterface 2/24/2019

The Team ! Jordi Adell (UPC, Barcelona) Ana Huerta (T.U. Madrid) Irene Kotsia (A.U. Thessaloniki) Benoit Macq (UCL, Belgium) Olivier Martin (UCL, Belgium) Hannes Pirker (OFAI, Vienna) Arman Savran (Boun, Istanbul) Rafaël Sebbe (TCTS, Mons) [Alexandre Benoît(INPG, Grenoble)] 2/24/2019