Compensatory coarticulation, /u/-fronting, and sound change in Standard Southern British: an acoustic and perceptual study.* Jonathan Harrington, Felicitas.

Slides:



Advertisements
Similar presentations
Tom Lentz (slides Ivana Brasileiro)
Advertisements

Tone perception and production by Cantonese-speaking and English- speaking L2 learners of Mandarin Chinese Yen-Chen Hao Indiana University.
{ “Age” Effects on Second Language Acquisition Examination of 4 hypotheses related to age and language learning
SPPA 403 Speech Science1 Unit 3 outline The Vocal Tract (VT) Source-Filter Theory of Speech Production Capturing Speech Dynamics The Vowels The Diphthongs.
Plasticity, exemplars, and the perceptual equivalence of ‘defective’ and non-defective /r/ realisations Rachael-Anne Knight & Mark J. Jones.
Philip Harrison J P French Associates & Department of Language & Linguistic Science, York University IAFPA 2006 Annual Conference Göteborg, Sweden Variability.
Vowel Formants in a Spectogram Nural Akbayir, Kim Brodziak, Sabuha Erdogan.
Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
Synthesizing naturally produced tokens Melissa Baese-Berk SoundLab 12 April 2009.
Effects of Competence, Exposure, and Linguistic Backgrounds on Accurate Production of English Pure Vowels by Native Japanese and Mandarin Speakers Malcolm.
Unravelling variation and change in the short vowel system of RP Anne Fabricius SCALPS Research Group Roskilde University Sociolinguistics Symposium 16.
CS 551/651: Structure of Spoken Language Lecture 12: Tests of Human Speech Perception John-Paul Hosom Fall 2008.
Coarticulation Analysis of Dysarthric Speech Xiaochuan Niu, advised by Jan van Santen.
Vowels (again) February 23, 2010 The News For Thursday: Give me a (one paragraph or so) description of what you’re thinking of doing for a term project.
ENG 528: Language Change Research Seminar Sociophonetics: An Introduction Chapter 5: Vowels (Continued) Lindblom (1963), Undershoot.
Infant sensitivity to distributional information can affect phonetic discrimination Jessica Maye, Janet F. Werker, LouAnn Gerken A brief article from Cognition.
Speech perception 2 Perceptual organization of speech.
“Speech and the Hearing-Impaired Child: Theory and Practice” Ch. 13 Vowels and Diphthongs –Vowels are formed when sound produced at the glottal source.
Nuclear Accent Shape and the Perception of Prominence Rachael-Anne Knight Prosody and Pragmatics 15 th November 2003.
Speech and speaker normalization (in vowel normalization)
AS14.1 – The Role of Neural Oscillations in Auditory Resolution and Speech Perception – Implications for CAPD, Phonological Awareness and Reading Disorders.
Do Children Pick and Choose? An Examination of Phonological Selection and Avoidance in Early Lexical Acquisition. Richard G. Schwartz and Laurence B. Leonard.
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Development of coarticulatory patterns in spontaneous speech Melinda Fricke Keith Johnson University of California, Berkeley.
TEMPLATE DESIGN © Perceptual compensation for /u/-fronting in American English KATAOKA, Reiko Department.
TEMPLATE DESIGN © Listener’s variation in phoneme category boundary as a source of sound change: a case of /u/-fronting.
SPEECH PERCEPTION The Speech Stimulus Perceiving Phonemes Top-Down Processing Is Speech Special?
What is Phonetics? Short answer: The study of speech sounds in all their aspects. Phonetics is about describing speech. (Note: phonetics ¹ phonics) Phonetic.
GABRIELLA RUIZ LING 620 OHIO UNIVERSITY Cross-language perceptual assimilation of French and German front rounded vowels by novice American listeners and.
5. 1 JPEG “ JPEG ” is Joint Photographic Experts Group. compresses pictures which don't have sharp changes e.g. landscape pictures. May lose some of the.
Distinguishing Evidence Accumulation from Response Bias in Categorical Decision-Making Vincent P. Ferrera 1,2, Jack Grinband 1,2, Quan Xiao 1,2, Joy Hirsch.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Present Experiment Introduction Coarticulatory Timing and Lexical Effects on Vowel Nasalization in English: an Aerodynamic Study Jason Bishop University.
Sebastián-Gallés, N. & Bosch, L. (2009) Developmental shift in the discrimination of vowel contrasts in bilingual infants: is the distributional account.
Diphthongs Five most frequent diphthongs in Māori are /ai ae au ou ao/. Mergers between /ai~ae/ and /au~ou/ [3] (Figure 2). Only one of these mergers is.
Segmental factors in language proficiency: Velarization degree as a signature of pronunciation talent Henrike Baumotte and Grzegorz Dogil {henrike.baumotte,
The Albertan æ/ɛ shift and community grammars
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
Jonathan Harrington Contextual ambiguities in speech signals and their consequences for sound change.
Dr. Richard Young Optronic Laboratories, Inc..  Uncertainty budgets are a growing requirement of measurements.  Multiple measurements are generally.
Perceived prominence and nuclear accent shape Rachael-Anne Knight LAGB 5 th September 2003.
Acoustic Phonetics 3/9/00. Acoustic Theory of Speech Production Modeling the vocal tract –Modeling= the construction of some replica of the actual physical.
Adaptive Design of Speech Sound Systems Randy Diehl In collaboration with Bjőrn Lindblom, Carl Creeger, Lori Holt, and Andrew Lotto.
1 PATTERN COMPARISON TECHNIQUES Test Pattern:Reference Pattern:
SPM short course – Oct Linear Models and Contrasts Jean-Baptiste Poline Neurospin, I2BM, CEA Saclay, France.
Understanding Action Verbs- Embodied Verbal Semantics Approach Pavan Kumar Srungaram M.Phil Cognitive Science (09CCHL02) Supervisor: Prof. Bapi.
Sh s Children with CIs produce ‘s’ with a lower spectral peak than their peers with NH, but both groups of children produce ‘sh’ similarly [1]. This effect.
Pragmatically-guided perceptual learning Tanya Kraljic, Arty Samuel, Susan Brennan Adaptation Project mini-Conference, May 7, 2007.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
SEPARATION OF CO-OCCURRING SYLLABLES: SEQUENTIAL AND SIMULTANEOUS GROUPING or CAN SCHEMATA OVERRULE PRIMITIVE GROUPING CUES IN SPEECH PERCEPTION? William.
Phonetic Context Effects Major Theories of Speech Perception Motor Theory: Specialized module (later version) represents speech sounds in terms of intended.
To examine the feasibility of using confusion matrices from speech recognition tests to identify impaired channels, impairments in this study were simulated.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Syllables and Stress October 21, 2015.
0 / 27 John-Paul Hosom 1 Alexander Kain Brian O. Bush Towards the Recovery of Targets from Coarticulated Speech for Automatic Speech Recognition Center.
Perceptual distance & sound change GSAS workshop on historical linguistics Oct
Language Perception.
Nuclear Accent Shape and the Perception of Syllable Pitch Rachael-Anne Knight LAGB 16 April 2003.
Transitions + Perception March 25, 2010 Tidbits Mystery spectrogram #3 is now up and ready for review! Final project ideas.
AUDITORY CORTEX 1 SEPT 11, 2015 – DAY 8 Brain & Language LING NSCI Fall 2015.
Bridging the gap between L2 speech perception research and phonological theory Paola Escudero & Paul Boersma (March 2002) Presented by Paola Escudero.
A STUDY ON PERCEPTUAL COMPENSATION FOR / /- FRONTING IN A MERICAN E NGLISH Reiko Kataoka February 14, 2009 BLS 35.
A PRODUCTION STUDY ON PHONOLOGIZATION OF /U/-FRONTING IN ALVEOLAR CONTEXT Reiko Kataoka 10 January 2009 LSA annual meeting.
Speech Perception in Infants Peter D. Eimas, Einar R. Siqueland, Peter Jusczyk, and James Vigorito 1971.
/u/-fronting in RP: a link between sound change and diminished perceptual compensation for coarticulation? Jonathan Harrington, Felicitas Kleber, Ulrich.
Does the brain compute confidence estimates about decisions?
Usage-Based Phonology Anna Nordenskjöld Bergman. Usage-Based Phonology overall approach What is the overall approach taken by this theory? summarize How.
K Nearest Neighbor Classification
Ala’a Spaih Abeer Abu-Hantash Directed by Dr.Allam Mousa
Uma R. Karmarkar, Dean V. Buonomano  Neuron 
Presentation transcript:

Compensatory coarticulation, /u/-fronting, and sound change in Standard Southern British: an acoustic and perceptual study.* Jonathan Harrington, Felicitas Kleber, Ulrich Reubold * submitted Journal of the Acoustical Society of America

General aim of this paper To establish to what extent a sound-change in progress, /u/-fronting in Standard Southern British (SSB) – can be linked to diminished perceptual compensation for coarticulation in Ohala's (1993) model of sound change. Background A. Perceptual compensation for coarticulation B. Ohala's model of sound change C. /u/-fronting in SSB

A. What is perceptual compensation for coarticulation? si su Frequency of fricative noise ACOUSTICS 1. Anticipatory coarticulatory lip-rounding causes spectral centre of gravity lowering in /s/ PERCEPTION Listeners reverse the effects of coarticulation sisu 2. Listeners know this and reverse its effect (= compensation for coarticulation*) *e.g. Fujisaki & Kunisaki, 1977; Mann & Repp, 1980 ;

If you synthesise a continuum from /s/ to /S/ that can be done by spectral COG lowering and prepend it to a vowel, then listeners are more likely to perceive the SAME synthetic token as /s/ before /u/ than before /a/ Perceptual compensation for coarticulation s S Frequency of noise

+ /u/ + /a/ Listener compensates for coarticulation (= factors out COG lowering assumed to be attributable to /u/): s s s s S s s s s S S S s S

B. Cental ideas in Ohala's model of sound change Ohala: "Today's variability is tomorrow's sound change" The origin of many sound changes is not always in the mouth of the speaker, but in the ear of the listener Contra sociolinguists: sound change is not teleological = it's not done on purpose or for any reason, it happens by accident because of an unintended error on the part of the listener (and also for this reason, the origin of sound change is not cognitive nor phonological) Hypoarticulation-induced sound change: one that arises out of the natural processes of coarticulation and in which the listener fails to compensate for coarticulation…

(the listener thinks: "the speaker meant to say /ci/") Hypoarticulation-induced sound change in Ohala Listener as speakerSpeakerListener /c/ has been phonologised because it is planned, produced and perceived, even in contexts that can't be explained by coarticulation plans /ki/ [ci] produces [ci] acoustic signal compensates for coarticulation reconstructs /ki/ Sound change: /k/ -> /c/ plans /ci, cu/ [ci, cu] reconstructs /ci/

C. /u/-fronting in Standard Southern British Extensive auditory and some acoustic evidence that SSB /u/ has fronted in the last 50 years e.g., Gimson, 1966, Wells 1982, Henton 1983, Deterding 1997, Hawkins & Midgley 2005, Roach, 1997)‏. ([i] ‘heed’, [] ‘hoard’, [u], ‘who’d’) /u/-fronting and possible chain-shifting in the Queen's Christmas broadcasts (Harrington, 2008)* Harrington (2008), Laboratory Phonology IX, in press

This could be just such an example of a hypoarticulation- induced sound change. Why? 1. Taking into account word-frequency, /u/ frequently (p ≈ 0.7) occurs in a coarticulatory fronting context (e.g. 'you', 'too', 'lose', 'do', 'new'). j u  F2-locus target distance plans /ju/ [j  ] produces F2 Time Speaker

2. In an analysis of the Christmas broadcasts at 20 year intervals, the Queen, Harrington (2008)* shows just such a reduction of F2 locus-target distance: locus-target distance Decade * Harrington (2008), Laboratory Phonology 9 in press

Our extension of Ohala's model to these data and age-differences in SSB speakers is as follows: /u/-fronting and speech perception Acoustic input: [s  n] compensate for coarticulation Perceived as: /sun/ /s  n/ OLD listenersYOUNG listeners

/u/-fronting, speech perception and production PRODUCTION PERCEPTION /fj  d/ /f  d/ [f  d] sound change Front Back Young Old For the Old, the allophones diverge in production, but not in perception (this is the trigger for sound change) For the Young, the allophones are aligned in perception and production and NO compensation for coarticulation [fj  d] [fud] /fjud/ /fud/ compensate for coarticulation (food) (feud)

[fj  d] [fud] PRODUCTION PERCEPTION /fjud/ /fud/ /fj  d/ /f  d/ [f  d] sound changecompensate for coarticulation Front Back Young Old Predictions about age differences 1. (trivially) /u/ vowels are fronted for the Young. 2. C-on-/u/ coarticulation is greater in the Old (their /u/ allophones show greater divergence). Production 3. Young and Old differ primarily on the back allophones (if sound change involves a shift of these to the front). Perception 4. The Old but not the Young compensate perceptually for coarticulation.

Experimental analysis I: Production (Predictions 1 -3)

30 Standard Southern British speakers recruited through University of Cambridge and University College London. YOUNG: 14 subjects aged (11 F, 3 M)‏ OLD: 13 subjects aged (7 F, 9 M) Method: Speakers Subjects were carefully checked to ensure that they were SSB speakers. only 1 subject took part in the production study only

Materials Isolated word production of words each produced 10 times /u/ C(C) Word  jused269 fjfeud266 hjhewed266 kjqueued270 ffood269 ssoup270 kcooed264 hwho'd270 swswoop268  2412 /i/ C(C) Word  jyeast269 ffeed267 hheed270 kkeyed270 sseep269 swsweep269  1614 / A / C(C) Word  hhard270 Recordings made in U.K. with SpeechRecorder (Draxler & Jänsch, 2004)

Acoustic parameters Formants calculated and F2 was checked and manually corrected. Each F2-trajectory was reduced to a single point in a three- dimensional space formed from the first three coefficients of the discrete-cosine-transformation (Watson & Harrington, 1999; Harrington, 2006; Zahorian, and Jagharghi, 1993) We did this because we wanted to assess vowel fronting in the entire F2-trajectory (from onset to offset) rather than just at the vowel's temporal midpoint (which encodes no dynamic information). Also with the DCT, we avoid having to make an often arbitrary decision about the location of the vowel target.

Discrete-cosine-transformation decomposes any signal into a set of ½ cycle cosine waves which, if summed, reconstruct entirely the original signal. The amplitudes of these cosine waves are the DCT coefficients. Moreover, the cosine waves at the lowest frequencies encode important properties of the trajectory's shape… …at frequency (rad/sample)… …is proportional to the trajectory's: DCT-coeff DCT-0average mean linear slope curvature 0 DCT-11 DCT-22

so you can use this technique to smooth formants… AnalysisAnalysis DCT coeffs But the important point for this paper is that each F2 trajectory is reduced to a single point in a 3D-space which encodes a smoothed trajectory, like the one on the right. F2 (Hz) SynthesisSynthesis

For each speaker separately, we quantified the extent of /u/-fronting in this DCT space by calculating each [u] token's relative distance to the front and back /i/ and /A/ vowel centroids Quantification of /u/-fronting d u = log(E 1 /E 2 ) = log(E 1 ) – log(E 2 ) Log Euclidean distance ratio d u = 0, [u] equidistant bet. /i/ and /A/ d u < 0, [u] nearer /A/ (back) d u > 0, [u] nearer /i/ (front) So if following hypothesis 1 /u/ is phonetically more back in the Old, then d u should be lower for the Old compared with the Young

2. Quantification of C-on-/u/ perseverative coarticulation We measured separately for each speaker the Inter- Euclidean distance in the DCT space between 'swoop' and 'used' swoop ju used wu DCT-0 DCT-1 swoop tokens to used centroid used tokens to swoop centroid If the coarticulatory influences of C-on-/u/ are greater in Older speakers (= hypothesis 2), then 'swoop' (/w/ has a backing influence) and 'used' (/j/ has a fronting influence) will be further apart i.e., the distances between them will be greater than for the Young.

Hypothesis 3: the age-difference is context-specific i.e., if sound change involves a shift of back allophones towards the front, then Young and Old should differ more on words with a non-front allophone ('food') than those with a front allophone ('feud') We calculated the distance in the DCT-space between Old and Young speakers together separately for each word. Thus the prediction is that the distance between e.g. Old/Young 'food' is expected to be greater than between Old/Young 'feud'.

Results 1: Young speakers have a fronter /u/ F1 x F2 plots at vowel midpoint

(When d u = 0, /u/ is equidistant between /i/ and /A/) Log Euclidean distance ratio, d u Front Back Results 1: Young speakers have a fronter /u/

Results 2: a greater C-on-/u/ influence for the Old Averaged, linearly time-normalised F2 trajectories in used and swoop

Euclidean distance between 'used' and 'swoop' Results 2: a greater C-on-/u/ influence for the Old

Results 3: smaller age difference for words where /u/ is in a fronting context C has a fronting effect on /u/? No Yes Euclidean distance between Young and Old separately per word (i.e., the sound change involves a shift of back allophones to the front)

Part II: Speech Perception

Method: synthetic continua We used HLSYN to create two 13 step synthetic /i-u/ continua at equal Bark intervals by varying F2 in two sets of minimal pairs : (a)/jist/ --- /just/YEAST---USED (p. tense) (b)/swip/ --- /swup/SWEEP---SWOOP A separate group of listeners verified that the endpoints of the continua could be correctly identified. Stimuli randomised and both continua presented in one session 5 times (5 x 13 x 2 = 130 randomised stimuli). Forced-choice identification task: Subjects responded with one of ''used'', ''yeast'', ''swoop'', ''sweep'' to each stimulus.

Speech perception predictions F2 highF2 low u i F2 highF2 low u i F2 highF2 low u i OLD left-shift relative to Old because they have a fronter /u/ and no (or much less) compensation for coarticulation YOUNG yeast- used boundary sweep- swoop boundary

so the different predicted responses are: F2 high F2 low u i u i 1. WORD: yeast-used boundary left-shifted relative to sweep-swoop (red vs blue) 2. AGE: Young left-shifted relative to Old (dash vs. straight) 3. AGE x WORD Small difference between Young vs. Old on yeast-used (red dash vs. red straight), big difference between Young vs. Old on sweep-swoop (blue dash vs. blue straight).

Results 1: WORD Significantly greater proportion of /u/ responses (across both age groups) in YEAST-USED relative to SWEEP- SWOOP (compatibly with Mann & Repp, 1980).

Results (2)‏: AGE The /i-u/ boundary is significantly left-shifted (greater proportion of /u/ responses) in YOUNG compared with OLD speakers.

Prediction 3. Small difference between YOUNG vs. OLD on yeast-used … big difference between YOUNG vs. OLD on sweep-swoop YOUNGOLD Results (3)‏: WORD x AGE

In the s. The large phonetic separation between allophones of /u/ in production required coarticulatory compensation to realign them as the same category in perception. For whatever reason, young listeners give up on compensating for coarticulation. They do not attribute [  ] to context but presume that /  / (not /u/) was intended by the speaker. Thus, /  / becomes phonologised... …leading to [  ] in their own productions even in contexts like 'food', 'move' where it has no coarticulatory explanation. Discussion I: /u/-fronting and sound- change in terms of Ohala's model.

But which comes first? Ohala: first give up on compensating for coarticulation, then there is a realignment in speech production (= sound change): loss of coarticulatory compensation in perception results in sound change in production. Or perhaps: first there is sound change and then there is a loss of perceptual compensation for coarticulation? This story would be compatible with exemplar theory (Pierrehumbert, 2003) as follows:

1. /u/ in Standard British English occurs most often (about 70% of the time) after consonants with a high F2 locus ('you', 'too', 'lose', 'do', 'new') 2. According to exemplar theory, this imbalance in lexical and phonological frequency is the trigger for sound change – i.e., the infrequently occurring back [u] in low F2-locus contexts (e.g., 'move', 'swoop') is perceptually unstable and so will be inclined to shift towards the more frequently occurring front allophone [  ]. 3. As this shift occurs in production, there will be less need to compensate for coarticulation in perception, because the allophones will be in progressively closer alignment. So this story has: sound change first, then loss of perceptual compensation. /u/-fronting, sound change, exemplar theory

A way forward may be to investigate compensatory coarticulation in young children. Perhaps children learn to compensate for coarticulation relatively late (Ohala, 1993) – perhaps they are more inclined than adults to perceive (incorrectly) an allophone as intended by the speaker. We might then have an alternative explanation for why sound change is so often led by the young. According to the sociolinguistics, this is intentional: the young lead sound change because they want to sound different from their parents/elder generation. From the Queen to language acquisition But in terms of Ohala (1993) it would be unintentional and come about because children are just not as good at compensating for coarticulation perceptually…

So as usual we need another experiment… Thank you for listening!