TEMPLATE DESIGN © 2008 Perceptual compensation for /u/-fronting in American English KATAOKA, Reiko Department.

Slides:



Advertisements
Similar presentations
Tom Lentz (slides Ivana Brasileiro)
Advertisements

Tone perception and production by Cantonese-speaking and English- speaking L2 learners of Mandarin Chinese Yen-Chen Hao Indiana University.
Sounds that “move” Diphthongs, glides and liquids.
Plasticity, exemplars, and the perceptual equivalence of ‘defective’ and non-defective /r/ realisations Rachael-Anne Knight & Mark J. Jones.
Speech Perception Dynamics of Speech
Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
CS 551/651: Structure of Spoken Language Lecture 12: Tests of Human Speech Perception John-Paul Hosom Fall 2008.
Acoustic Characteristics of Vowels
Coarticulation Analysis of Dysarthric Speech Xiaochuan Niu, advised by Jan van Santen.
ENG 528: Language Change Research Seminar Sociophonetics: An Introduction Chapter 5: Vowels (Continued) Lindblom (1963), Undershoot.
Phonetic variability of the Greek rhotic sound Mary Baltazani University of Ioannina, Greece  Rhotics exhibit considerable phonetic variety cross-linguistically.
Infant sensitivity to distributional information can affect phonetic discrimination Jessica Maye, Janet F. Werker, LouAnn Gerken A brief article from Cognition.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Ling 240: Language and Mind Acquisition of Phonology.
Speech perception 2 Perceptual organization of speech.
Splice: From vowel offset to vowel onset FIG 3. Example of stimulus spliced from the repetitive syllables. EXPERIMENT 2 (Voicing ID) METHOD Speech materials:
“Speech and the Hearing-Impaired Child: Theory and Practice” Ch. 13 Vowels and Diphthongs –Vowels are formed when sound produced at the glottal source.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Speech and speaker normalization (in vowel normalization)
Perception of syllable prominence by listeners with and without competence in the tested language Anders Eriksson 1, Esther Grabe 2 & Hartmut Traunmüller.
PHONETICS AND PHONOLOGY
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Spectrogram & its reading
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Cognitive Processes PSY 334 Chapter 2 – Perception April 9, 2003.
TEMPLATE DESIGN © Listener’s variation in phoneme category boundary as a source of sound change: a case of /u/-fronting.
PSY 369: Psycholinguistics
SPEECH PERCEPTION The Speech Stimulus Perceiving Phonemes Top-Down Processing Is Speech Special?
Speech Perception Richard Wright Linguistics 453.
TEMPLATE DESIGN © A Production Study on Phonologization of /u/-fronting in Alveolar Context KATAOKA, Reiko
What is Phonetics? Short answer: The study of speech sounds in all their aspects. Phonetics is about describing speech. (Note: phonetics ¹ phonics) Phonetic.
Phonetics, day 2 Oct 3, 2008 Phonetics 1.Experimental a. production b. perception 2. Surveys/Interviews.
Stop Place Contrasts before Liquids Edward Flemming MIT.
The Perception of Speech
Cognitive Processes PSY 334 Chapter 2 – Perception.
Speech Perception. Phoneme - a basic unit of a speech sound that distinguishes one word from another Phonemes do not have meaning on their own but they.
Experiments concerning boundary tone perception in German 3 rd Workshop of the SPP-1234 Potsdam, 7 th January 2009 Presentation of the Stuttgart Project.
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
Infant Speech Perception & Language Processing. Languages of the World Similar and Different on many features Similarities –Arbitrary mapping of sound.
1 Speech Perception 3/30/00. 2 Speech Perception How do we perceive speech? –Multifaceted process –Not fully understood –Models & theories attempt to.
Speech Perception1 Fricatives and Affricates We will be looking at acoustic cues in terms of … –Manner –Place –voicing.
Adaptive Design of Speech Sound Systems Randy Diehl In collaboration with Bjőrn Lindblom, Carl Creeger, Lori Holt, and Andrew Lotto.
Results 1.Boundary shift Japanese vs. English perceptions Korean vs. English perceptions 1.Category boundary was shifted toward boundaries in listeners’
Results Tone study: Accuracy and error rates (percentage lower than 10% is omitted) Consonant study: Accuracy and error rates 3aSCb5. The categorical nature.
Sh s Children with CIs produce ‘s’ with a lower spectral peak than their peers with NH, but both groups of children produce ‘sh’ similarly [1]. This effect.
Epenthetic vowels in Japanese: a perceptual illusion? Emmanual Dupoux, et al (1999) By Carl O’Toole.
Sensation & Perception
Phonetic Context Effects Major Theories of Speech Perception Motor Theory: Specialized module (later version) represents speech sounds in terms of intended.
The New Normal: Goodness Judgments of Non-Invariant Speech Julia Drouin, Speech, Language and Hearing Sciences & Psychology, Dr.
Acoustic Continua and Phonetic Categories Frequency - Tones.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Neurophysiologic correlates of cross-language phonetic perception LING 7912 Professor Nina Kazanina.
Chapter 13: Speech Perception. The Acoustic Signal Produced by air that is pushed up from the lungs through the vocal cords and into the vocal tract Vowels.
Katherine Morrow, Sarah Williams, and Chang Liu Department of Communication Sciences and Disorders The University of Texas at Austin, Austin, TX
Bosch & Sebastián-Gallés Simultaneous Bilingualism and the Perception of a Language-Specific Vowel Contrast in the First Year of Life.
Speech Perception.
Language Perception.
WebCT You will find a link to WebCT under the “Current Students” heading on It is your responsibility to know how to work WebCT!
Transitions + Perception March 25, 2010 Tidbits Mystery spectrogram #3 is now up and ready for review! Final project ideas.
Acoustic Phonetics 3/14/00.
A STUDY ON PERCEPTUAL COMPENSATION FOR / /- FRONTING IN A MERICAN E NGLISH Reiko Kataoka February 14, 2009 BLS 35.
A PRODUCTION STUDY ON PHONOLOGIZATION OF /U/-FRONTING IN ALVEOLAR CONTEXT Reiko Kataoka 10 January 2009 LSA annual meeting.
/u/-fronting in RP: a link between sound change and diminished perceptual compensation for coarticulation? Jonathan Harrington, Felicitas Kleber, Ulrich.
Danielle Werle Undergraduate Thesis Intelligibility and the Carrier Phrase Effect in Sinewave Speech.
4aPPa32. How Susceptibility To Noise Varies Across Speech Frequencies
Cognitive Processes PSY 334
What is Phonetics? Short answer: The study of speech sounds in all their aspects. Phonetics is about describing speech. (Note: phonetics ¹ phonics) Phonetic.
Speech Perception.
Volume 64, Issue 3, Pages (November 2009)
Speech Communications
Presentation transcript:

TEMPLATE DESIGN © Perceptual compensation for /u/-fronting in American English KATAOKA, Reiko Department of Linguistics, University of California at Berkeley, 1203 Dwinelle Hall, Berkeley, California Objectives Summary and DiscussionMaterials and Methods (cont.)Results: Experiment 2 Citation Introduction Listener's identification of speech sounds are influenced by both perceived and expected characteristics due to the influence of surrounding sounds. For example, a vowel ambiguous between /i/ and /e/ is heard more often as /e/ when the precursor sentence has low F1 but it is heard as /i/ when the precursor has high F1 (Ladefoged & Broadbent 1957), and a greater degree of vowel undershoot is perceptually accepted in fast speech than in slow speech (Lindblom & Studdert-Kennedy 1967). Later, Ohala and Feder (1994) showed that American listeners judge a vowel stimulus which is ambiguous between /i/ and /u/ more frequently as /u/ in alveolar context than in bilabial context, and do so both when the context is heard and when it is “restored”. The current study is an attempt to extend Ohala & Feder’s study with additional measures to reveal the locus of perceptual compensation in the human speech processing system. Materials and Methods (cont.) Presentation: A stimulus was played after a precursor “I guess the word is…”. Each stimulus was tested 5 times. Fillers ([CiC], [C ɨ C], [CuC], and [C ʉ C], each three times) were mixed in the trials to create genuine vowel quality variation. Test stimulus and fillers were presented at random order. During a trial, the two alternative words (/CiC/ and /CuC/ forms) appear on the screen. (See Table 1.) The words on the screen will make subjects to ‘restore’ the onset and coda phoneme even when hearing the stimulus with white noise. Task : Two-alternative forced-choice. The subjects determined which of the two words on the screen the word just heard was. Subjects: 31 native speakers of American English (10-M, 21-F) Factors: 1)Consonantal context (3 conditions: Alveolar, Bilabial, Zero) 2)Status of context (2 conditions: Acoustic or Restored) Table 1. Factors tested, stimuli presented (e.g. [dyt]), and two alternatives shown on the screen for each stimulus. EXPERIMENT 2 Methods are same as Experiment 1 except that precursor was uttered in 3 different speech rates. Subjects: Same as Experiment 1. Factors: Precursor speech rate (3 conditions: fast, medium, slow). EXPERIMENT 3 Production Subjects uttered a sentence “Say who again.” 10 times. From the mid point of the vowel in ‘who’, F2 value was measured. Then, by using each subject’s other vowel productions (/i, ɪ, e, æ, a, ɔ, ʌ, ʊ, u/, 10 times), normalized mean F2 value for /u/ was obtained as a difference from mean F2 of all the vowels on the log scale (LN). Perception Table 3. # of responses by Context and Condition (5 repetition X 31 listeners = 155 cell total) Results: Experiment 3 To replicate Ohala & Feder’s findings and further investigate the context effect on Reaction Time (RT) (Experiment 1) To test the effect of speech rate on the degree of perceptual compensation (Experiment 2) To investigate the relationship between speech production and perceptual compensation (Experiment 3) Materials and Methods EXPERIMENT 1 Stimuli: Re-synthesized syllables [dyt], [byp], and [y] (to be used in Acoustic condition) and another set where white noise replaced the consonants (to be used in Restored condition) were created by: Iterating single vowel period in the CV transition of ‘dude’ to obtain a steady vowel (i.e. no formant transition) of 100 ms Adding amplitude contour for the first and last 15 ms of the vowel Adding F0 contour: 130 at onset  90 Hz at offset Adding excised natural stop burst (/b/ or /d/) at vowel onset and another (/p/ or /t/) 70 ms after the vowel offset, or(Fig. 1, a) Adding white noise in the place of natural stop burst(Fig. 1, b) (a) (b) Figure 1. Waveforms of a /dVt/ stimulus used in Acoustic Context condition (a) and Restored Context condition (b). Stimuli: 3 continua of 10 syllables each, varying by vowel backness (i.e., /dit/ - /dut/, /bip/ - /bup/, and /i/ - /u/ continuum) were created by the following process: Create 10 equal-step /i/ - /u/ continuum. The voice source was extracted from a sustained vowel, which was produced by the speaker of a precursor sentence, by applying an inverse filtering of LPC object of the vowel so that the voice of the stimuli matches to the voice of the precursor. All 10 vowels have identical formant structures except for F2 value, which varied from 2,300 Hz (step 1) to 860 Hz (step 10). See Fig. 2. Add amplitude contour and F0 contour as in Exp.1. Add onset and coda consonant bursts as in Exp.1. Presentation and Task: Same as Experiment 1 except that each stimulus was tested 4 times. Subjects: Same as Experiment 1. Factor: Consonantal context (3 conditions: Alveolar, Bilabial, Zero). Figure 2. Spectrograms of 10-step /i/ - /u/ continuum. Formant frequency (Hz) for the vowels are: F1=375, F2 = variable, F3 = 2500, F4 = 3500, and F5 = Results: Experiment 1 Table 2. # of responses by Context and Condition (5 repetition X 31 listeners = 155 cell total) Chi-Square test for association: # of /u/-response and Contexts Acoustic Condition [χ 2 =3.429 (2), p=0.180] Restored Condition [χ 2 =0.612 (2), p=0.736] Figure 3. Reaction Time for /u/ response. Effect of Contexts Restored: [F=4.5 (1.5, 43.3), p<0.05] Acoustic: [F=3.0 (1.8, 52.9), p=0.06] Figure 4. Reaction Time for /u/ response. Figure 5. % /u/-response as a function of stimulus number, in three different consonant conditions. Context StatusAlveolarBilabialZero Acoustic[dyt][byp][y] Restored*[NyN] Alternatives‘deet’ or ‘doot’‘beep’ or ‘boop’‘ee’ or ‘oo’ Cond.Resp. Context Total G. Total AlveolarBilabialZero Restored/i/ /u/ Acoustic/i/ /u/ Total Cond.Resp. Context Total G. Total AlveolarBilabialZero Fast/i/ /u/ Med./i/ /u/ Slow/i/ /u/ Total Figure 6. Scatter plot of the perceptual /u/ space against normalized F2 of /u/. Figure 7. Metric for obtaining the perceptual space for the vowel /u/. The space is defined as an area under the % /u/-response function. Harrington, J., Kleber, F., and Reubold, U. ( 2008 ). “Compensation for coarticulation, /u/-fronting, and sound change in standard southern British: An acoustic and perceptual study,” J. Acoust. Soc. Am. 123, Ladefoged, P., and Broadbent, D. E. ( 1957 ). “Information conveyed by vowels,” J. Acoust. Soc. Am. 29, Lindblom, B., and Studdert-Kennedy, M. ( 1967 ). “On the role of formant transitions in vowel recognition,” J. Acoust Soc. Am. 42, Ohala, J. J., and Feder, D. ( 1994 ). “Listeners’ identification of speech sounds is influenced by adjacent ‘restored’ phonemes,” Phonetica 51, Listeners heard an ambiguous stimulus more often as /u/ in Alveolar than in Bilabial or Zero context, and did so both in Acoustic and Restored phoneme conditions (Experiment 1). Also, the /u/ identification function shifted as a function of consonantal context (Experiment 3). These results confirm listener’s ability to perceptually compensate for coarticulatory fronting of /u/ in alveolar contexts by shifting the category boundary. It also seems to show listeners’ ability to utilize mentally stored acoustic images of linguistic contexts to perform compensation. Alveolar context shortens the RT for /u/-identification (Experiment 1 & 2). This suggests that the perceptual system not only shifts the perceptual category boundary but also to facilitate perceptual processing for a category that is contrastive to the context. Speech rate of precursor influenced vowel identification in Bilabial and Zero context (Experiment 2), suggesting the listener’s ability to perform not only categorical but also gradient compensation. The effect was absent in Alveolar context. This could be due to ceiling effect in this particular experimental setting. This effect needs to be re-examined by using an improved method. Mild correlation was found between /u/ identification in perception and F2 measured from /u/ production (Experiment 3). Among those who produced /u/ with relatively low F2 (back /u/), many also accepted fronted /u/s as category members in perceptual vowel categorization task. Although previous studies (e.g. Harrington et. al., 2008) demonstrated robust link between production and perception, language users seem to be much more tolerant to speech variation (e.g. dialects, sociolects, style shifts, etc.) that differs from their own production patterns. Effect of Context Fast: [F=6.0 (2, 48), p<0.05] Med: [F=9.3 (2, 52), p<0.001] Slow: [F=10.2 (2, 60), p<0.001]