Subjective evaluation of an emotional speech database for Basque Aholab Signal Processing Laboratory – University of the Basque Country Authors: I. Sainz,

Slides:



Advertisements
Similar presentations
1 Using the HTK speech recogniser to analyse prosody in a corpus of German spoken learners English Toshifumi Oba, Eric Atwell University of Leeds, School.
Advertisements

The Role of F0 in the Perceived Accentedness of L2 Speech Mary Grantham O’Brien Stephen Winters GLAC-15, Banff, Alberta May 1, 2009.
HARMONIC MODEL FOR FEMALE VOICE EMOTIONAL SYNTHESIS Anna PŘIBILOVÁ Department of Radioelectronics, Slovak University of Technology Ilkovičova 3, SK-812.
Sub-Project I Prosody, Tones and Text-To-Speech Synthesis Sin-Horng Chen (PI), Chiu-yu Tseng (Co-PI), Yih-Ru Wang (Co-PI), Yuan-Fu Liao (Co-PI), Lin-shan.
Chapter 3 Generalization: How broadly do the results apply?
Associations of behavioral parameters of speech emotional prosody perception with EI measures in adult listeners Elena Dmitrieva Kira Zaitseva, Alexandr.
Emotional Speech Analysis using Artificial Neural Networks IMCSIT-AAIA October 18-20, 2010 – Wisla, Poland. Jana Tuckova & Martin Sramka Department of.
Acoustic Model Adaptation Based On Pronunciation Variability Analysis For Non-Native Speech Recognition Yoo Rhee Oh, Jae Sam Yoon, and Hong Kook Kim Dept.
ACCENT MODIFICATION OF POLYSYLLABIC WORDS IN A PHRASE Marina Moscow – подготовка речи и презентации.
Can Non-Native English Speakers Detect and Identify Native English Speakers’ Dialectal Variations? Rebecca Austerman.
Facial expression as an input annotation modality for affective speech-to-speech translation Éva Székely, Zeeshan Ahmed, Ingmar Steiner, Julie Carson-Berndsen.
Vocal Emotion Recognition with Cochlear Implants Xin Luo, Qian-Jie Fu, John J. Galvin III Presentation By Archie Archibong.
Dr. O. Dakkak & Dr. N. Ghneim: HIAST M. Abu-Zleikha & S. Al-Moubyed: IT fac., Damascus U. Prosodic Feature Introduction and Emotion Incorporation in an.
Me First!: The Use of Personal Pronouns in Canadian English Katarina Fischer, Jenny Kenmir, Natalie Poirier Queen’s University Hypotheses 1.Native Canadian.
Human Language Technologies – Text-to-Speech © 2007 IBM Corporation Sixth Speech Synthesis Workshop, Bonn, Germany.August 22-24, 2007 Automatic Exploration.
MULTI LINGUAL ISSUES IN SPEECH SYNTHESIS AND RECOGNITION IN INDIAN LANGUAGES NIXON PATEL Bhrigus Inc Multilingual & International Speech.
On the use of Intonation in ASR: preliminary results March, 13th 14th 2003 Meeting COST 275 Halmstad.
Emotion in Meetings: Hot Spots and Laughter. Corpus used ICSI Meeting Corpus – 75 unscripted, naturally occurring meetings on scientific topics – 71 hours.
EMOTIONS NATURE EVALUATION BASED ON SEGMENTAL INFORMATION BASED ON PROSODIC INFORMATION AUTOMATIC CLASSIFICATION EXPERIMENTS RESYNTHESIS VOICE PERCEPTUAL.
Comparing feature sets for acted and spontaneous speech in view of automatic emotion recognition Thurid Vogt, Elisabeth André ICME 2005 Multimedia concepts.
Advanced Technology Center Stuttgart EMOTIONAL SPACE IMPROVES EMOTION RECOGNITION Raquel Tato, Rocio Santos, Ralf Kompe Man Machine Interface Lab Advance.
1 Evidence of Emotion Julia Hirschberg
Hypothesis testing simplified. Objective: Describe a variable - Mean, Median, Mode, Standard Deviation Examples: What is the percentage of totally satisfied.
Independent Sample T-test Often used with experimental designs N subjects are randomly assigned to two groups (Control * Treatment). After treatment, the.
A Text-to-Speech Synthesis System
Numerical Text-to-Speech Synthesis System Presentation By: Sevakula Rahul Kumar.
Lecture 14 Testing a Hypothesis about Two Independent Means.
STANDARDIZATION OF SPEECH CORPUS Li Ai-jun, Yin Zhi-gang Phonetics Laboratory, Institute of Linguistics, Chinese Academy of Social Sciences.
Occasion:HUMAINE / WP4 / Workshop "From Signals to Signs of Emotion and Vice Versa" Santorin / Fira, 18th – 22nd September, 2004 Talk: Ronald Müller Speech.
Utterance Verification for Spontaneous Mandarin Speech Keyword Spotting Liu Xin, BinXi Wang Presenter: Kai-Wun Shih No.306, P.O. Box 1001,ZhengZhou,450002,
Prepared by: Waleed Mohamed Azmy Under Supervision:
Automated Scoring of Picture- based Story Narration Swapna Somasundaran Chong Min Lee Martin Chodorow Xinhao Wang.
Machine Learning in Spoken Language Processing Lecture 21 Spoken Language Processing Prof. Andrew Rosenberg.
Copyright 2007, Toshiba Corporation. How (not) to Select Your Voice Corpus: Random Selection vs. Phonologically Balanced Tanya Lambert, Norbert Braunschweiler,
On Speaker-Specific Prosodic Models for Automatic Dialog Act Segmentation of Multi-Party Meetings Jáchym Kolář 1,2 Elizabeth Shriberg 1,3 Yang Liu 1,4.
Synthesis of Child Speech With HMM Adaptation and Voice Conversion Oliver Watts, Junichi Yamagishi, Member, IEEE, Simon King, Senior Member, IEEE, and.
SPEECH CONTENT Spanish Expressive Voices: Corpus for Emotion Research in Spanish R. Barra-Chicote 1, J. M. Montero 1, J. Macias-Guarasa 2, S. Lufti 1,
Building a sentential model for automatic prosody evaluation Kyuchul Yoon School of English Language & Literature Yeungnam University Korea.
Information Technology – Dialogue Systems Ulm University (Germany) Speech Data Corpus for Verbal Intelligence Estimation.
Speech Perception 4/4/00.
A prosodically sensitive diphone synthesis system for Korean Kyuchul Yoon Linguistics Department The Ohio State University.
Segmental encoding of prosodic categories: A perception study through speech synthesis Kyuchul Yoon, Mary Beckman & Chris Brew.
Evaluating prosody prediction in synthesis with respect to Modern Greek prenuclear accents Elisabeth Chorianopoulou MSc in Speech and Language Processing.
Text independent speaker identification in multilingual environments I. Luengo, E. Navas, I. Sainz, I. Saratxaga, J. Sanchez, I. Odriozola and I. Hernaez.
COPYRIGHT © All rights reserved by Sound acoustics Germany The averaged quality measures over all test cases indicate the real influence of a test object.
Handing Uncertain Observations in Unsupervised Topic-Mixture Language Model Adaptation Ekapol Chuangsuwanich 1, Shinji Watanabe 2, Takaaki Hori 2, Tomoharu.
Basic Strategies for Speaking English Speaking English is a demanding skill in ELT. Learners who are not native speaker of English are forced to speak.
Imposing native speakers’ prosody on non-native speakers’ utterances: Preliminary studies Kyuchul Yoon Spring 2006 NAELL The Division of English Kyungnam.
1/17/20161 Emotion in Meetings: Business and Personal Julia Hirschberg CS 4995/6998.
Introduction A field survey of Dutch language resources has been carried out within the framework of a project launched by the Dutch Language Union (Nederlandse.
Can a blind person guess the state of mind of someone they are talking with without seeing them? SAK-WERNICKA, JOLANTA. "EXPLORING THEORY OF MIND USE IN.
Integrating Multiple Knowledge Sources For Improved Speech Understanding Sherif Abdou, Michael Scordilis Department of Electrical and Computer Engineering,
Automatic Pronunciation Scoring of Specific Phone Segments for Language Instruction EuroSpeech 1997 Authors: Y. Kim, H. Franco, L. Neumeyer Presenter:
St. Petersburg Institute for Informatics and Automation of the Russian Academy of Sciences Recurrent Neural Network-based Language Modeling for an Automatic.
1 7-Speech Recognition Speech Recognition Concepts Speech Recognition Approaches Recognition Theories Bayse Rule Simple Language Model P(A|W) Network Types.
Acoustic Cues to Emotional Speech Julia Hirschberg (joint work with Jennifer Venditti and Jackson Liscombe) Columbia University 26 June 2003.
Interpreting Ambiguous Emotional Expressions Speech Analysis and Interpretation Laboratory ACII 2009.
Flexible Speaker Adaptation using Maximum Likelihood Linear Regression Authors: C. J. Leggetter P. C. Woodland Presenter: 陳亮宇 Proc. ARPA Spoken Language.
Acoustic phonetics: Concerned with describing the acoustics of speech. Also called speech acoustics. Big questions: (1) What are the relationships between.
G. Anushiya Rachel Project Officer
The Foreign Language Effect on The Sunk-Cost Fallacy
Towards Emotion Prediction in Spoken Tutoring Dialogues
Course Projects Speech Recognition Spring 1386
3.0 Map of Subject Areas.
Copyright © American Speech-Language-Hearing Association
TITLE Authors Institution RESULTS INTRODUCTION CONCLUSION AIMS METHODS
Emotional Speech Julia Hirschberg CS /8/2018.
Informational Text.
EMOTIONS & FEELINGS Communicative aim: Talk about emotions / express feelings. Language objective: Study words related with emotions and feelings.
Within-speaker variability in long-term F0
Presentation transcript:

Subjective evaluation of an emotional speech database for Basque Aholab Signal Processing Laboratory – University of the Basque Country Authors: I. Sainz, I. Saratxaga, E. Navas, I. Hernáez, J. Sanchez, I. Luengo, I. Odriozola

Index  Introduction  Corpus Design and Recording  Evaluation Process  Results  Conclusions & Work in Progress

Introduction Speech Synthesis  Intelligibility  Naturalness  Variability Emotional Speech  Prosody  Acoustic Features  Corpus Based Tecniques  Implicit Acoustic modeling

Corpus Design and Recording Aim  Emotional Corpus-based TTS  Prosodic and acoustic analysis Recording emotional speech  Spontaneous  Elicited  Acted 2 speakers, 702 sentences  1 hour per emotion

Corpus Design and Recording Big 6 Emotions

Evaluation Process Subjetive evaluation campaign  30 stimuli per actor  Forced choice 20 Subjects  11 native speakers

Results Confusion Matrix Recognition average 76.6%  Far above chance level 17%

Results Confusion Matrices(Actress & Actor) Average 75.83% Average 76.50%

Results Effects of Listeners Student’s t-test  Determine if the hypothesis is true Groups (recognition rate)  Women (72.78%) Vs Men (77.62%)  Not significant (95% confidence interval)  Native speakers (77.12%) Vs Non-native (75%)  Not significant (95% confidence interval)  1 st Half of the test (72.64%) Vs 2 nd Half (79.7%)  t=2.85  p= < 0.05 Significant!  Recognition rate increase of 7% –Almost constant for all groups

Conclusions & Work in progress Valid resource of emotional speech for Basque language  Emotions readily recognized for both actors Study and modeling of emotional speech Corpus based TTS with emotions  Neutral already developed

Subjective evaluation of an emotional speech database for Basque Aholab Signal Processing Laboratory – University of the Basque Country Authors: I. Sainz, I. Saratxaga, E. Navas, I. Hernáez, J. Sanchez, I. Luengo, I. Odriozola