Download presentation
Presentation is loading. Please wait.
Published byJoseph Stone Modified over 8 years ago
1
Towards Semantic Affect Sensing in Sentences Alexander Osherenko
2
Goal –Language independent approach for affect sensing in textual corpora containing spontaneous emotional dialogues Method –Extracting features and evaluating resulting datasets by standard data mining approaches considering language independence
3
Overview Properties of classified corpora Feature extraction Results: –SAL –AIBO –SmartKom Conclusions Outlook
4
Properties of classified dialogues Corpora may be in different languages No obvious signs of emotional meaning. Utterances are short. Can be grammatically incorrect, contain repairs, repetitions and inexact wordings. Can convey contradicting emotional meaning. Utterances are interdependent (can be seen as a continuous stream of information).
5
Feature extraction Most frequent (stemmed) utterance words in the current corpus (in most cases only seventh/eighth of the whole list) History as most frequent (stemmed) words in the current and n previous utterances (ditto) No dependence on an affect words‘ list e.g. Whissell‘s dictionary of affect
6
Dialogue corpora SAL –QUB (Cowie 2006) AIBO –Univ. of Erlangen (Batliner et.al. 2004) SmartKom –Univ. of Munich (Steininger et al. 2002)
7
SAL Instance – utterance in transliteration 670 in FEELTRACE annotated utterances Agreement – 65.2778% 3 affect states English corpus FEELTRACE scores (mapped onto classes pos./neutral/neg.)
8
Evaluation for a three class problem in SAL SMO in WEKA Cross-validation (10 fold) Overall number of words - 1486 rev.precisionrecallfMeasure#wordshistory maj.74.9972.1373.474034 cc79.0171.6574.851854 dr66.5262.9764.557432 em72.9870.0271.362124 jd70.0262.8565.911854
9
AIBO Instance – paragraph in transliteration 3990 instances Sparse transliteration texts (commands to AIBO) 4 affect states German corpus
10
Evaluation for a four class problem in AIBO SMO in WEKA Learning/testing sets (1738/2252 resp. 2252/1738) Only words’, not history features Overall number of words – 488 (!) precisionrecallfMeasure#wordshistory 53.3349.0050.1595 39.3930.2227.19955
11
SmartKom Wizard of Oz scenario Instance – turn 817 annotated instances 11 user states German corpus
12
Evaluation for n class problem in SmartKom SMO in WEKA Cross-validation (10 fold) Overall number of words - 1125 different affect states#whistoryPRfM joyful- strong joyful- weak surprisedneutralhelplessangry- weak angry- strong 562543.2838.1640.16 joyfulsurprisedneutralhelplessangry562647.0542.9944.61 joyfulneutralhelplessangry1125545.4945.4545.39 joyfulneutralproblem1125655.5354.5554.93 no problemhelplessangry562554.6451.7752.99 no problemproblem562569.7667.1468.31 not angryangry281572.8266.6869.16
13
Conclusions Higher number of words and longer history don’t induce better classification, rather their combination Extracted features can serve as a basis (AIBO results – sparse data, repetitious content) Erroneous classification could have been caused by the discrepancy between the rating and the corresponding text Language-independent features
14
Outlook Further feature extraction (combination, history of POS groups?) Studying erroneous instances (esp. in SMARTKOM) Multimodality (prosodic/lexical) Application for journalistic articles e.g. movie reviews Is 100% precision the goal?
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.