Human language technologies Data Collections & Studies WP4- Emotions: Faces.

Slides:



Advertisements
Similar presentations
1 Preparing Future multiSensorial inTerAction Research – human language technologies PF-STAR – the project. HUMAINE WP5-WS, Belfast, Dec.
Advertisements

APPROACHES TO T&L Language
Juli 2010 eNTERFACE Application : Surveillance in trains Video, Audio processing Sound localization, pattern rec.
HUMAINE Summer School - September Basic Emotions from Body Movements HUMAINE Summer School 2006 Casa Paganini Genova, Italy Ahmad S. Shaarani The.
I Can See It All Over Your Face
Audio-based Emotion Recognition for Advanced Information Retrieval in Judicial Domain ICT4JUSTICE 2008 – Thessaloniki,October 24 G. Arosio, E. Fersini,
Matthias Wimmer, Ursula Zucker and Bernd Radig Chair for Image Understanding Computer Science Technische Universität München { wimmerm, zucker, radig
Collection and Analysis of Multimodal Interaction in Direction Giving Dialogues Seikei University Takeo TsukamotoYumi Muroya Masashi Okamoto Yukiko Nakano.
Facial expression as an input annotation modality for affective speech-to-speech translation Éva Székely, Zeeshan Ahmed, Ingmar Steiner, Julie Carson-Berndsen.
 INTRODUCTION  STEPS OF GESTURE RECOGNITION  TRACKING TECHNOLOGIES  SPEECH WITH GESTURE  APPLICATIONS.
Languages & The Media, 5 Nov 2004, Berlin 1 New Markets, New Trends The technology side Stelios Piperidis
Double level analysis of the Multimodal Expressions of Emotions in Human-Machine Interaction.
1 Affective Learning with an EEG Approach Xiaowei Li School of Information Science and Engineering, Lanzhou University, Lanzhou, China
Coye Cheshire & Andrew Fiore June 13, 2015 // Computer-Mediated Communication Visual Interfaces for CMC.
A Social and Emotional Approach to Games and Human Computer Interaction (HCI) Katherine Isbister Associate Professor, Digital Media and CSE.
MUSCLE movie data base is a multimodal movie corpus collected to develop content- based multimedia processing like: - speaker clustering - speaker turn.
1 Evidence of Emotion Julia Hirschberg
1 A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions Zhihong Zeng, Maja Pantic, Glenn I. Roisman, Thomas S. Huang Reported.
EKMAN’S FACIAL EXPRESSIONS STUDY A Demonstration.
1 IUT de Montreuil Université Paris 8 Emotion in Interaction: Embodied Conversational Agents Catherine Pelachaud.
Producing Emotional Speech Thanks to Gabriel Schubiner.
Recognizing Emotions in Facial Expressions
Sunee Holland University of South Australia School of Computer and Information Science Supervisor: Dr G Stewart Von Itzstein.
Body Language I don’t know. Flying kiss to show love Thumb up: Great! Fist up to show his objection.
Building the Design Studio of the Future Aaron Adler Jacob Eisenstein Michael Oltmans Lisa Guttentag Randall Davis October 23, 2004.
WOZ acoustic data collection for interactive TV A. Brutti*, L. Cristoforetti*, W. Kellermann+, L. Marquardt+, M. Omologo* * Fondazione Bruno Kessler (FBK)
Scripting: Practicing Verbal Interaction Chapter 33, p Idalia Gannon Brenda Ayala Lewis EDC S382S- ESL Methods Summer 2010.
Laboratory of Computational Engineering Michael Frydrych, Making the head smile Smile? -> part of non-verbal communication Functions of non-verbal.
GUI: Specifying Complete User Interaction Soft computing Laboratory Yonsei University October 25, 2004.
Writing Movie Reviews. Pair Activity While watching the video, answer the following questions on a size 2: How did the two critics begin their review.
Multi-modal expression of Swedish prominence Björn Granström Centre for Speech Technology, Department of Speech, Music and Hearing, KTH, Stockholm, Sweden.
The persuasive import of gesture and gaze. The importance of bodily behaviour in persuasive discourse has been acknowledged as early as in ancient Rome.
NM – LREC 2008 /1 N. Moreau 1, D. Mostefa 1, R. Stiefelhagen 2, S. Burger 3, K. Choukri 1 1 ELDA, 2 UKA-ISL, 3 CMU s:
SPEECH CONTENT Spanish Expressive Voices: Corpus for Emotion Research in Spanish R. Barra-Chicote 1, J. M. Montero 1, J. Macias-Guarasa 2, S. Lufti 1,
1 Ying-li Tian, Member, IEEE, Takeo Kanade, Fellow, IEEE, and Jeffrey F. Cohn, Member, IEEE Presenter: I-Chung Hung Advisor: Dr. Yen-Ting Chen Date:
Nespole!’s Experiment on Multimodality (Summer 2001) Erica Costantini (University of Trieste) Fabio Pianesi (ITC-irst, Trento) Susanne Burger (CMU)
Perceptual Analysis of Talking Avatar Head Movements: A Quantitative Perspective Xiaohan Ma, Binh H. Le, and Zhigang Deng Department of Computer Science.
The Expression of Emotion: Nonverbal Communication.
1 Workshop « Multimodal Corpora » Jean-Claude MARTIN Patrizia PAGGIO Peter KÜEHNLEIN Rainer STIEFELHAGEN Fabio PIANESI.
Feedback Elisabetta Bevacqua, Dirk Heylen,, Catherine Pelachaud, Isabella Poggi, Marc Schröder.
Multimodality, universals, natural interaction… and some other stories… Kostas Karpouzis & Stefanos Kollias ICCS/NTUA HUMAINE WP4.
Intelligent Control and Automation, WCICA 2008.
Steve Oliver English UK South West Conference Saturday October 24 th 2015.
Performance Comparison of Speaker and Emotion Recognition
Getting Data ● Kinds of data for linguistics – Written – Spoken – Visual (ASL, body language) ● Phonetics – Implosives-larynx lowering, rounding, x-ray.
MIT Artificial Intelligence Laboratory — Research Directions The Next Generation of Robots? Rodney Brooks.
Electronic visualization laboratory, university of illinois at chicago Towards Lifelike Interfaces That Learn Jason Leigh, Andrew Johnson, Luc Renambot,
Jennifer Lee Final Automated Detection of Human Emotion.
Ekman’s Facial Expressions Study A Demonstration.
Facial Expressions and Emotions Mental Health. Total Participants Adults (30+ years old)328 Adults (30+ years old) Adolescents (13-19 years old)118 Adolescents.
“Articulatory Talking Head” Showcase Project, INRIA, KTH. Articulatory Talking Head driven by Automatic Speech Recognition INRIA, Parole Team KTH, Centre.
Presented By Meet Shah. Goal  Automatically predicting the respondent’s reactions (accept or reject) to offers during face to face negotiation by analyzing.
Interpreting Ambiguous Emotional Expressions Speech Analysis and Interpretation Laboratory ACII 2009.
Recognition and Expression of Emotions by a Symbiotic Android Head Daniele Mazzei, Abolfazl Zaraki, Nicole Lazzeri and Danilo De Rossi Presentation by:
Day 8 Usability testing.
Under Guidance of Mr. A. S. Jalal Associate Professor Dept. of Computer Engineering and Applications GLA University, Mathura Presented by Dev Drume Agrawal.
What is it? Details Look at the whole.  Form of communication ◦ Communicating without words ◦ Non-verbal ◦ Using facial expressions ◦ Gestures  Can.
Nataliya Nadtoka James Edge, Philip Jackson, Adrian Hilton CVSSP Centre for Vision, Speech & Signal Processing UNIVERSITY OF SURREY.
Applications · E-learning apps, changing the presentation style of an e-lesson according to the mood of the person. · Psychological health services, to.
2D-CSF Models of Visual System Development. Model of Adult Spatial Vision.
Automated Detection of Human Emotion
External Communication
Communication Skills COMM 101 Lecture#2
Warm Up Get out your notes/flash cards and study your vocabulary words (act, cast, character, comedy, complication, drama, improvisation, lines, monologue,
What is blue eyes ? aims on creating computational machines that have perceptual and sensory ability like those of human beings. interactive computer.
Séance 1 HOW ARE YOU ?.
Presented by: Mónica Domínguez
COMMUNICATION.
Automated Detection of Human Emotion
A Generative Audio-Visual Prosodic Model for Virtual
Presentation transcript:

human language technologies Data Collections & Studies WP4- Emotions: Faces

Data Collections & Studies WP4 – Emotions: Faces Collection and annotation of audio-visual databases extensive data collection, both at KTH and ISTC/IRST – using opto-electronic systems reflective markers placed on the subject’s face capturing of dynamics of emotional facial expressions with very high precision. eliciting technique: using movies to elicit facial expressions denoting emotions on watching subjects  attempted – not promising extraction technique: extract expressive behaviour directly from movies and television talk-shows  attempted – not promising

Data Collections & Studies WP4 – Emotions: Faces KTH - first year DATABASE 1 15 emotions and attitudes were recorded (acted) anger, fear, surprise, sadness, disgust, happiness, worry, satisfaction, insecurity, confidence, questioning, encouragment, doubt, confirmation and neutral Semantically neutral utterances, 9 utterances per expression DATABASE 2 6 emotional states confident, confirming, questioning, insecure, happy, neutral VCV & VCCV nonsense words CVC nonsense words Short sentences Common ITA-SWE set (abba, adda, alla, anna, avva) DATABASE 3 Spontaneous dialogue

Data Collections & Studies WP4 – Emotions: Faces Collection of audio-visual databases: interactive dialogues (KTH)  Eliciting technique: information seeking scenario  Focus on the speaker who has the role of information giver  The speaker whose facial and head motion is to be recorded seats facing 4 infrared cameras, a digital video- camera,a microphone and his/her interlocutor.

Data Collections & Studies WP4 – Emotions: Faces ISTC & IRST - first year 6 emotional states (Ekman’s set) + Neutral – Anger, Disgust, Fear, Happiness, Sadness, Surprise – 3 intensities (Low, Medium, High) “isolated” emotional expressions VCV nonsense words (aba, ada, aLA, adZa, ala, ana, ava) – good phonetic coverage of Italian Long sentence (“Il fabbro lavora con forza usando il martello e la tenaglia” – “the smith works with strength using the hammer and the pincer”) common ITA-SWE set (VCCV nonsense words: abba, adda, alla, anna, avva) “concatenated” emotional expressions VCV nonsense words, in pairs, with different emotions e.g. (aba) Neutral – (aba) Happy

Data Collections & Studies WP4 – Emotions: Faces Results ISTC/IRST: 1573 recordings – 798 single emotional expressions (7 emotional states, 3 intensities – L, M, H) – 672 concatenated emotional expressions (in pairs, 3 emotional states - Anger, Happy, Neutral - medium intensity) – 57 long sentences (7 emotional states, 3 intensities) – 46 instances of the common ITA-SWE set (3 emotional states, medium intensity) KTH: 1749 recordings (database 2) – 828 VCV words (138 x 6 emotional states) – 246 CVC words (41 x 6 emotional states) – 645 sentences (270 neutral + 75 x 5 emotional states) – 30 instances of the common ITA-SWE set Total: 3322 recordings

Data Collections & Studies WP4 – Emotions: Faces Qualisys recordings: Swedish db – 2nd year 75 sentences with Ekman’s 6 basic emotions + neutral Dialogues to analyze communicative facial expressions: – 10 short dialogues in a travel agency scenario 15 sentences uttered with a focussed word, with the 6 expressions used in corpus 2 + anger Example: Båten seglade förbi Båten seglade förbi

Data Collections & Studies WP4 – Emotions: Faces Short videos containing acted kinetic facial expressions (video length: 4-27 secs.) 8 professional actors (4 male and 4 female). Each actor played Ekman’s set of six emotions (happy, sad, angry, surprised, disgusted, afraid) + neutral Actors were asked to play each emotion on three intensity levels (Low -Medium – High) Total: 1008 short videos (= ~ 2h 50’) Audio-Visual Italian Database – 2nd year (IRST) : A database of human facial expressions (1) See Poster!!!

Data Collections & Studies WP4 – Emotions: Faces A database of human facial expressions (2)  Facial expressions recorded in two conditions: 1. “utterance” condition: actors played emotions while uttering a phonetically rich and visemically balanced sentence. 2. “non utterance” condition: actors played emotions without pronouncing any sentence.  Both video and audio signals were recorded.  After collecting the corpus: Data Selection – Validation of the emotions played – Video selection based on the accordance among judges. “In quella piccola stanza vuota c’era pero’ soltanto una sveglia”,, See Poster!!!,

Data Collections & Studies WP4 – Emotions: Faces Annotation of audio-visual databases: interactive dialogues ANVIL: tool for the analysis of digitized audio-visual data Orthographic transcription of the dialogue Annotation of the facial expressions related to emotions and of the communicative gestures (turn-taking, feedback and so on) The annotation is performed on a freely definable multi- layered annotation scheme, created ad hoc for the specific purposes. These levels go from a less detailed to a more detailed analysis Annotation is performed on several main tracks, which are displayed, on the screen in alignment with the video and audio data

Data Collections & Studies WP4 – Emotions: Faces Annotation (cont’d) glad

Data Collections & Studies WP4 – Emotions: Faces Evaluation Studies (IRST) Experiment 1: Comparison of emotion recognition rates from natural (actor) videos with different types of synthetic (synthetic face) videos, in different animation conditions [reference person: Fabio Pianesi – Experiment 2: Cross-cultural comparison of emotion recognition rates from Italian and Swedish natural and synthetic videos [reference person: Fabio Pianesi – Experiment 3: as for Experiment 1 but using – three regions of the face – only one animation condition (script based) [reference person: Michela Prete –

Data Collections & Studies WP4 – Emotions: Faces Papers on Evaluation Studies J. Beskow, L. Cerrato, P. Cosi, E. Costantini, M. Nordstrand, F. Pianesi, M. Prete, G. Svanfeldt, "Preliminary Cross-cultural Evaluation of Expressiveness in Synthetic Faces". In E. André, L. Dybkiaer, W. Minker, P. Heisterkamp (eds.) "Affective Dialogue Systems", ADS '04, Springer Verlag. Berlin, E. Costantini, F. Pianesi, P. Cosi, "Evaluation of Synthetic Faces: Human Recognition of Emotional Facial Displays ". In E. André, L. Dybkiaer, W. Minker, P. Heisterkamp (eds.) "Affective Dialogue Systems". Springer Verlag, Berlin, 2004 E. Costantini, F. Pianesi, M. Prete "Recognising Emotions in Human and Synthetic Faces: The Role of the Upper and Lower Parts of the Face". To appear in Proceedings of IUI 2005: International Conference on Intelligent User Interfaces. San Diego, California, 2005.