Towards Emotion Prediction in Spoken Tutoring Dialogues

Slides:



Advertisements
Similar presentations
Mihai Rotaru Diane J. Litman DoD Group Meeting Presentation
Advertisements

Detecting Certainness in Spoken Tutorial Dialogues Liscombe, Hirschberg & Venditti Using System and User Performance Features to Improve Emotion Detection.
Uncertainty Corpus: Resource to Study User Affect in Complex Spoken Dialogue Systems Kate Forbes-Riley, Diane Litman, Scott Silliman, Amruta Purandare.
Spoken Language Technologies: A review of application areas and research issues Analysis and synthesis of F0 contours Agnieszka Wagner Department of Phonetics,
Student simulation and evaluation DOD meeting Hua Ai 03/03/2006.
Spoken Language Processing Lab Who we are: Julia Hirschberg, Stefan Benus, Fadi Biadsy, Frank Enos, Agus Gravano, Jackson Liscombe, Sameer Maskey, Andrew.
Outline Why study emotional speech?
Using Emotion Recognition and Dialog Analysis to Detect Trouble in Communication in Spoken Dialog Systems Nathan Imse Kelly Peterson.
Detecting missrecognitions Predicting with prosody.
Classification of Discourse Functions of Affirmative Words in Spoken Dialogue Julia Agustín Gravano, Stefan Benus, Julia Hirschberg Shira Mitchell, Ilia.
Topics = Domain-Specific Concepts Online Physics Encyclopedia ‘Eric Weisstein's World of Physics’ Contains total 3040 terms including multi-word concepts.
Annotating Student Emotional States in Spoken Tutoring Dialogues Diane Litman and Kate Forbes-Riley Learning Research and Development Center and Computer.
Predicting Student Emotions in Computer-Human Tutoring Dialogues Diane J. Litman and Kate Forbes-Riley University of Pittsburgh Pittsburgh, PA USA.
® Automatic Scoring of Children's Read-Aloud Text Passages and Word Lists Klaus Zechner, John Sabatini and Lei Chen Educational Testing Service.
Modeling User Satisfaction and Student Learning in a Spoken Dialogue Tutoring System with Generic, Tutoring, and User Affect Parameters Kate Forbes-Riley.
Short Introduction to Machine Learning Instructor: Rada Mihalcea.
Interactive Dialogue Systems Professor Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh Pittsburgh,
circle A Comparison of Tutor and Student Behavior in Speech Versus Text Based Tutoring Carolyn P. Rosé, Diane Litman, Dumisizwe Bhembe, Kate Forbes, Scott.
Kate’s Ongoing Work on Uncertainty Adaptation in ITSPOKE.
Spoken Dialogue for Intelligent Tutoring Systems: Opportunities and Challenges Diane Litman Computer Science Department & Learning Research & Development.
circle Adding Spoken Dialogue to a Text-Based Tutorial Dialogue System Diane J. Litman Learning Research and Development Center & Computer Science Department.
Comparing Synthesized versus Pre-Recorded Tutor Speech in an Intelligent Tutoring Spoken Dialogue System Kate Forbes-Riley and Diane Litman and Scott Silliman.
Adaptive Spoken Dialogue Systems & Computational Linguistics Diane J. Litman Dept. of Computer Science & Learning Research and Development Center University.
Correlations with Learning in Spoken Tutoring Dialogues Diane Litman Learning Research and Development Center and Computer Science Department University.
Experiments with ITSPOKE: An Intelligent Tutoring Spoken Dialogue System Dr. Diane Litman Associate Professor, Computer Science Department and Research.
A Successful Dialogue without Adaptation S: Hi, this is AT&T Amtrak schedule system. This is Toot. How may I help you? U: I want a train from Baltimore.
Collaborative Research: Monitoring Student State in Tutorial Spoken Dialogue Diane Litman Computer Science Department and Learning Research and Development.
Predicting Student Emotions in Computer-Human Tutoring Dialogues Diane J. Litman&Kate Forbes-Riley University of Pittsburgh Department of Computer Science.
Recognizing Discourse Structure: Speech Discourse & Dialogue CMSC October 11, 2006.
Why predict emotions? Feature granularity levels [1] uses pitch features computed at the word-level Offers a better approximation of the pitch contour.
Using Word-level Features to Better Predict Student Emotions during Spoken Tutoring Dialogues Mihai Rotaru Diane J. Litman Graduate Research Competition.
Diane Litman Learning Research & Development Center
Spoken Dialogue in Human and Computer Tutoring Diane Litman Learning Research and Development Center and Computer Science Department University of Pittsburgh.
Speech and Language Processing for Adaptive Training Diane Litman Professor, Computer Science Department Senior Scientist, Learning Research & Development.
Spoken Dialog Systems Diane J. Litman Professor, Computer Science Department.
Using Prosody to Recognize Student Emotions and Attitudes in Spoken Tutoring Dialogues Diane Litman Department of Computer Science and Learning Research.
(Speech and Affect in Intelligent Tutoring) Spoken Dialogue Systems Diane Litman Computer Science Department and Learning Research and Development Center.
Metacognition and Learning in Spoken Dialogue Computer Tutoring Kate Forbes-Riley and Diane Litman Learning Research and Development Center University.
circle Spoken Dialogue for the Why2 Intelligent Tutoring System Diane J. Litman Learning Research and Development Center & Computer Science Department.
Modeling Student Benefits from Illustrations and Graphs Michael Lipschultz Diane Litman Computer Science Department University of Pittsburgh.
circle Towards Spoken Dialogue Systems for Tutorial Applications Diane Litman Reprise of LRDC Board of Visitors Meeting, April 2003.
Spoken Dialogue for Intelligent Tutoring Systems: Opportunities and Challenges Diane Litman Computer Science Department Learning Research & Development.
Improving (Meta)cognitive Tutoring by Detecting and Responding to Uncertainty Diane Litman & Kate Forbes-Riley University of Pittsburgh Pittsburgh, PA.
Experiments with ITSPOKE: An Intelligent Tutoring Spoken Dialogue System Diane Litman Computer Science Department and Learning Research and Development.
User Simulation for Spoken Dialogue Systems Diane Litman Computer Science Department & Learning Research and Development Center University of Pittsburgh.
Acoustic Cues to Emotional Speech Julia Hirschberg (joint work with Jennifer Venditti and Jackson Liscombe) Columbia University 26 June 2003.
Using Natural Language Processing to Analyze Tutorial Dialogue Corpora Across Domains and Modalities Diane Litman, University of Pittsburgh, Pittsburgh,
Detecting and Adapting to Student Uncertainty in a Spoken Tutorial Dialogue System Diane Litman Computer Science Department & Learning Research & Development.
Prosodic Cues to Disengagement and Uncertainty in Physics Tutorial Dialogues Diane Litman, Heather Friedberg, Kate Forbes-Riley University of Pittsburgh.
Predicting and Adapting to Poor Speech Recognition in a Spoken Dialogue System Diane J. Litman AT&T Labs -- Research
Predicting Emotion in Spoken Dialogue from Multiple Knowledge Sources Kate Forbes-Riley and Diane Litman Learning Research and Development Center and Computer.
Language Identification and Part-of-Speech Tagging
Applications of Discourse Structure for Spoken Dialogue Systems
Neural Machine Translation
Linguistic knowledge for Speech recognition
Sentiment analysis algorithms and applications: A survey
Natural Language Processing for Enhancing Teaching and Learning
Chapter 6. Data Collection in a Wizard-of-Oz Experiment in Reinforcement Learning for Adaptive Dialogue Systems by: Rieser & Lemon. Course: Autonomous.
For Evaluating Dialog Error Conditions Based on Acoustic Information
Dialogue-Learning Correlations in Spoken Dialogue Tutoring
Temporal Argument Mining for Writing Assistance
Studying Intonation Julia Hirschberg CS /21/2018.
Detecting Prosody Improvement in Oral Rereading
Paradigms, Corpora, and Tools in Discourse and Dialog Research
Recognizing Structure: Sentence, Speaker, andTopic Segmentation
Searching and Summarizing Speech
Advanced NLP: Speech Research and Technologies
Introduction to Machine Translation
Recognizing Structure: Dialogue Acts and Segmentation
Low Level Cues to Emotion
Presentation transcript:

Towards Emotion Prediction in Spoken Tutoring Dialogues Diane Litman, Kate Forbes, and Scott Silliman Learning Research and Development Center and Computer Science Department University of Pittsburgh Pittsburgh, PA 15260 USA

Outline Introduction System and Corpora Pilot Study Summary

Motivation Human tutors listen to both “what” and “how” (e.g. “confident” vs. “uncertain”) Speech supplies acoustic-prosodic information about user state; some spoken dialogue applications already handle “problem” dialogues specially (Ang et al. 2002, Batliner et al. 2003, Litman et al. 2001)  Can effectiveness of computer dialogue tutors increase by detecting/adapting to student emotional state (Evens 2001)

ITSPOKE: Intelligent Tutoring SPOKEn Dialogue System “Back-end” is (text-based) Why2-Atlas intelligent tutoring dialogue system (VanLehn et al., 2002) Speech input via Sphinx2 speech recognizer Speech output via Festival text-to-speech synthesizer

ITSPOKE Screen Shot

Parallel Human-Human Corpus Target size - 20 subjects - up to 10 dialogues per subject Size on 5/20/03 - 10 subjects - 86 dialogues - 62 dialogues transcribed - 3066 manually segmented student turns

Human-Human Corpus Transcription and Annotation

Annotating Emotion 14 transcribed dialogues (n=553 student turns) Each student turn was annotated (intuition of 1 coder) with one of 3 general categories: -negative (e.g. ‘uncertain’ or ‘frustrated’): n=141 -positive (e.g. ‘confident’ or ‘certain’): n=167 -neutral/indeterminate: n=248 KVL: first bullet is confusing KVL: You could just put the numbers after the italic bullets above; clearer

Example Annotated Excerpt …6.5 minutes after essay… Tutor: Now this law that force is equal to mass times acceleration, what's this law called? This is uh since this it is a very important basic uh fact uh it is it is a law of physics. Um you have you have read it in the background material. Can you recall it? Student: Um no it was one of Newton's laws but I don't- remember which one. (laugh) (EMOTION = NEGATIVE) Tutor: Right, right- That- is Newton's second law of motion. Student: he I- Ok, because I remember one, two, and three, but I didn't know if there was a different name (EMOTION = POSITIVE) Tutor: Yeah that's right you know Newton was a genius- KVL: This text will not be visible on the 9th floor setup. How about just including the student turns and not the tutor turns on the slide?

Predicting Emotion Ripper (machine learning program) Input: 1) classes to be learned (our 3 emotion categories) 2) names and possible values for a set of features (next slide) 3) training examples with class and feature values (the annotated student turns) Output: an ordered set of if…then rules for classifying future examples

Pilot Machine Learning Results Six turn features - Problem, Student, Duration, StartTime, Transcription, #Words -all features automatically available in real-time Cross-validated error (33.03%) significantly lower than majority class baseline (55.69%) if (duration ≥ 0.65) & (text has “I”) then negative else if (duration ≥ 2.98) then negative else if (duration ≥ 0.93) & (startTime ≥ 297.62) then positive else if (text has “right”) then positive else neutral

Summary and Current Directions 1) Pilot study suggests there are indeed features that can be used to automatically predict emotion in tutoring dialogues Wider variety of features from many knowledge sources (e.g., pitch, amplitude, timing, other acoustic/prosodic, syntactic, semantic, discourse) Reliable Emotion Annotation Guidelines Analysis of Human-Computer Corpus 2) Empirical comparisons with typed tutorial dialogues (Building Educational Applications Using NLP paper)

Text-Feature Ruleset 1 Feature: Text in Turn Figure 2: Text-Feature Ruleset for Emotion Prediction (excerpt from 21 rules)) if (text has “the”) & (text has “don't”) then negative else if (text has “I”) & (text has “don't”) then negative … else if (text has “um”) & (text has “<hn>”) then negative else if (text has “the”) & (text has “<fs>”) then negative else if (text has “right”) then positive else if (text has “so”) then positive else if (text has “(laugh)”) & (text has “that's”) then positive else neutral Estimated mean error and standard deviation: 39.03% +/- 2.40%, based on 25-fold cross-validation