Two Types of Listeners? Marie Nilsenov á (Tilburg University) 1. Background When you and I listen to the same utterance, we may not perceive the linguistic.

Slides:



Advertisements
Similar presentations
Tone perception and production by Cantonese-speaking and English- speaking L2 learners of Mandarin Chinese Yen-Chen Hao Indiana University.
Advertisements

The Role of F0 in the Perceived Accentedness of L2 Speech Mary Grantham O’Brien Stephen Winters GLAC-15, Banff, Alberta May 1, 2009.
Hearing relative phases for two harmonic components D. Timothy Ives 1, H. Martin Reimann 2, Ralph van Dinther 1 and Roy D. Patterson 1 1. Introduction.
Auditory scene analysis 2
Voice quality variation with fundamental frequency in English and Mandarin.
The perception of dialect Julia Fischer-Weppler HS Speaker Characteristics Venice International University
Timbre perception. Objective Timbre perception and the physical properties of the sound on which it depends Formal definition: ‘that attribute of auditory.
Pitch Perception.
1 The Effect of Pitch Span on the Alignment of Intonational Peaks and Plateaux Rachael-Anne Knight University of Cambridge.
Chapter 4 How Music Works Part II: Pitch.
CHAPTER 12 Loudness and Pitch. Loudness/Pitch Loudness--psychological experience most directly related to sound pressure/intensity Pitch--psychological.
Music Perception. Why music perception? 1. Found in all cultures - listening to music is a universal activity. 2. Interesting from a developmental point.
Nuclear Accent Shape and the Perception of Prominence Rachael-Anne Knight Prosody and Pragmatics 15 th November 2003.
Vocal Emotion Recognition with Cochlear Implants Xin Luo, Qian-Jie Fu, John J. Galvin III Presentation By Archie Archibong.
AUTOMATIC SPEECH CLASSIFICATION TO FIVE EMOTIONAL STATES BASED ON GENDER INFORMATION ABSTRACT We report on the statistics of global prosodic features of.
A.Diederich– International University Bremen – USC – MMM – Spring 5 1 The Perception of Frequency cont'd.
A brief message from your TAs Tine Gulbrandsen Wahab Hanif.
On the Correlation between Energy and Pitch Accent in Read English Speech Andrew Rosenberg Weekly Speech Lab Talk 6/27/06.
Chapter three Phonology
The Science of Sound Chapter 8
Human Psychoacoustics shows ‘tuning’ for frequencies of speech If a tree falls in the forest and no one is there to hear it, will it make a sound?
Harmonics, Timbre & The Frequency Domain
Toshiba Update 04/09/2006 Data-Driven Prosody and Voice Quality Generation for Emotional Speech Zeynep Inanoglu & Steve Young Machine Intelligence Lab.
Fundamentals of Statistical Analysis DR. SUREJ P JOHN.
CNBH, Physiology Department, Cambridge University The perception of size in four families of instruments; brass, strings, woodwind.
Schizophrenia and Depression – Evidence in Speech Prosody Student: Yonatan Vaizman Advisor: Prof. Daphna Weinshall Joint work with Roie Kliper and Dr.
Categorizing Emotion in Spoken Language Janine K. Fitzpatrick and John Logan METHOD RESULTS We understand emotion through spoken language via two types.
Perceived prominence and nuclear accent shape Rachael-Anne Knight LAGB 5 th September 2003.
Exam 1 February 6 – 7 – 8 – 9 Moodle testing centre.
A study on Prediction on Listener Emotion in Speech for Medical Doctor Interface M.Kurematsu Faculty of Software and Information Science Iwate Prefectural.
The Effect of Cognitive Fatigue on Language in Multiple Sclerosis Leena Maria Heikkola, Päivi Hämäläinen and Juhani Ruutiainen Åbo Akademi University,
Critical Review on a Working Paper : Effects of background music, voice cues, earcons and gender on psychological ratings and heart rates during product.
A prosodically sensitive diphone synthesis system for Korean Kyuchul Yoon Linguistics Department The Ohio State University.
1. Background Evidence of phonetic perception during the first year of life: from language-universal listeners to native listeners: Consonants and vowels:
Evaluating prosody prediction in synthesis with respect to Modern Greek prenuclear accents Elisabeth Chorianopoulou MSc in Speech and Language Processing.
Phonetic Context Effects Major Theories of Speech Perception Motor Theory: Specialized module (later version) represents speech sounds in terms of intended.
1 Nonparametric Statistical Techniques Chapter 17.
The Effect of Pitch Span on Intonational Plateaux Rachael-Anne Knight University of Cambridge Speech Prosody 2002.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Hearing: Physiology and Psychoacoustics 9. The Function of Hearing The basics Nature of sound Anatomy and physiology of the auditory system How we perceive.
Effects of emotion induced by music on experimental pain Roy, Mathieu (1), Peretz, Isabelle (1), Rainville, Pierre (2) (1) Département de psychologie,
Performance Comparison of Speaker and Emotion Recognition
Introduction to psycho-acoustics: Some basic auditory attributes For audio demonstrations, click on any loudspeaker icons you see....
Nuclear Accent Shape and the Perception of Syllable Pitch Rachael-Anne Knight LAGB 16 April 2003.
What vocal cues indicate sarcasm? By: Jack Dolan Rockwell, P. (2000). Lower, slower, louder: Vocal cues of sarcasm. Journal of Psycholinguistic Research,
Timbre and Memory An experiment for the musical mind Emily Yang Yu Music 151, 2008.
Language and Speech, 2000, 43 (2), THE BEHAVIOUR OF H* AND L* UNDER VARIATIONS IN PITCH RANGE IN DUTCH RISING CONTOURS Carlos Gussenhoven and Toni.
Covert Attention Results (Spring). Covert Attention Results (Fall ’05)
Intersensory Redundancy Facilitates Infants’ Perception of Meaning in Speech Passages Irina Castellanos, Melissa Shuman, and Lorraine E. Bahrick Florida.
Music Or, a lecture to soothe the savage beast. Music Basics  What is music?  Sound and silence temporally organized  Sounds of music  Pitch, timbre,
Acoustic Cues to Emotional Speech Julia Hirschberg (joint work with Jennifer Venditti and Jackson Liscombe) Columbia University 26 June 2003.
Without Words for Emotions: Is the emotional processing deficit in alexithymia caused by dissociation or suppression? Christian Sinnott & Dr. Mei-Ching.
Unit 6 Unit 6 Teaching Pronunciation. Teaching aims able to understand the role of pronunciation in language learning able to know the goal of teaching.
Functions of Intonation By Cristina Koch. Intonation “Intonation is the melody or music of a language. It refers to the way the voice rises and falls.
RESEARCH MOTHODOLOGY SZRZ6014 Dr. Farzana Kabir Ahmad Taqiyah Khadijah Ghazali (814537) SENTIMENT ANALYSIS FOR VOICE OF THE CUSTOMER.
Building awareness and concern for pronunciation by Joanne Kenworthy - Teaching English Pronunciation FONETICA Y FONOLOGIA II - ALEXANDRA NAIR ZUÑIGA.
Presented By Meet Shah. Goal  Automatically predicting the respondent’s reactions (accept or reject) to offers during face to face negotiation by analyzing.
1 Nonparametric Statistical Techniques Chapter 18.
COMBINATION TONES The Science of Sound Chapter 8 MUSICAL ACOUSTICS.
Results Introduction The present study focuses on adult attitudes toward children. Many examples of discrimination against children in Western societies.
Zatorre paper Presented by MaryKate Chester
August 15, 2008, presented by Rio Akasaka
6th International Conference on Language Variation in Europe
Auditory Illusions Phy103 Physics of Music Fall 2005
Studying Intonation Julia Hirschberg CS /21/2018.
2 independent Groups Graziano & Raulin (1997).
Speech Perception CS4706.
Sam Norman-Haignere, Nancy G. Kanwisher, Josh H. McDermott  Neuron 
Sound, language, thought and sense integration
Music Signal Processing
Presentation transcript:

Two Types of Listeners? Marie Nilsenov á (Tilburg University) 1. Background When you and I listen to the same utterance, we may not perceive the linguistic (e.g., accenting, the asserting vs. questioning attitude) and paralinguistic (e.g., emotions) information in the signal in the same way. Specifically, we may differ in how much prominence we would assign to the complex pitch information included in the speech sound. As experiments on the perception of complex tones with a missing fundamental (F0) show, listeners can be categorized into two groups, depending on how they asses the relative pitch of a sound: F0-listeners pay attention to the virtual pitch calculated by their auditory system. (For each component of a complex tone, the system calculates missing lower harmonics. It looks for a subharmonic match - a detected prominence.) Spectral listeners focus on the energy at certain (low) levels in the signal, I.e., on the pitch of a single spectral component. With respect to the perception of music, F0- listeners seem to prefer short, sharp and impulsive tones (drums, guitar, piano, trumpet, flute); spectral listeners, on the other hand, tend to choose sustained tones and spectral information (lower-pitched strings, brass, organ or the signing voice). Schneider et al (2005) used MRI and MEG to look for a neural basis behind these two types of pitch perception and found that F0- listeners demonstrated a leftward asymmetry in the lateral Heschl’s gyrus, while spectral pitch listeners possessed the opposite asymmetry. So far, the bimodal distribution found among listeners has not been linked to the perception of information in speech stimuli, though a link to language learning has been established. In speech, information about the emotions and attitudes of the speaker can be conveyed both by F0 modulation, as well as by changes in the timbre. For example, F0 changes are typically associated with the speech act status of an utterance (question, assertion) and with the epistemic state of the speaker (certainty about the truth of the utterance), viz. Safarova (2006), while timbre has been linked to speaker’s emotions, especially milder affective states (e.g., tense voice and anger, breathy and creaky voice with boredom or sadness), viz. Gobl & Ni Chasaide (2003). We hypothesize that virtual pitch listeners would be more sensitive to the information contained in the F0 variation, while spectral pitch listeners would perceive changes in the spectrum as more salient. Research question: Are there systematical differences among listeners in terms of how they perceive information encoded in pitch? Hypothesis: F0-listeners judge speech stimuli more consistently on the basis of the perceived F0 changes than spectral listeners. 2. Methodology Pilot study A pilot study with 20 Dutch native speakers (10 musical professionals and 10 non- professionals, with an equal distribution of gender) and 12 complex-tone stimuli (each composed of two tones comprising two harmonics) confirmed the hypothesis. Participants who were identified as spectral pitch listeners were less consistent in their judgments on 16 speech-pitch pairs (with pitch extracted from the speech recording; 32 stimuli in total) for the presence of two emotions and two attitudes than participants who were virtual F0 listeners. Primary study An experimental study with two tasks: 1. Assessment of the listener type, based on a perception task with 36 synthesized complet two-tone stimuli (18 ambiguous and 18 control combinations). 2. Perception task with speech and sine- wave stimuli: 16 intonationally variable renderings of a complex sentence (‘Al draagt de aap een gouden ring het is en blijft een lelijk ding’) recorded by a female actor + 16 synthesized (hummed) pitch contours extracted from the speech stimuli. Participants: Thirty-four Dutch native speakers (10 male, 24 female) without professional musical training. Procedure: Participants first performed the listener- type categorical-judgment task during which they indicated whether they perceived a set of two-tone combinations to be falling or rising. Each combination was presented twice. Subsequently, they evaluated the speech and pitch-contour stimuli on a visual analogue scale for the presence/absence of the emotion ‘opgewekt’ (“happy”). 3. Conclusion & Discussion The listener type was calculated as  p = (f 0 - f sp )/(f 0 + f sp ); where f 0 = listener selected the virtual F0 melody f sp = listener selected the spectral melody. There was no bimodal distribution of  p ; instead, we found a non-normal distribution (M=.51, S.D.=.69; Shapiro-Wilk’s p=.001) of the score among our listeners with the listener type clearly skewed in the f 0 direction. Figure 1: The distribution of listener type, measured in the interval, with absolute spectral listeners scoring -1 and absolute virtual F0 listeners scoring 1, was skewed in the direction of F0-type listener. There was no (predicted negative) correlation between listener type and consistency score on the speech/pitch- contour VAS task (r=.036, p=.84). An assessment of the methodology employed in the experiment (for both tasks) is currently under way, as well as an additional testing in other languages. Contact Information Marie Nilsenov á Faculty of Arts PO BOX LE Tilburg The Netherlands Phone Number: +31(0)13 –