Interarticulator programming in VCV sequences: Effects of closure duration on lip and tongue coordination Anders Löfqvist Haskins Laboratories New Haven,

Slides:



Advertisements
Similar presentations
Motion in One Dimension
Advertisements

Sounds that “move” Diphthongs, glides and liquids.
SPPA 403 Speech Science1 Unit 3 outline The Vocal Tract (VT) Source-Filter Theory of Speech Production Capturing Speech Dynamics The Vowels The Diphthongs.
Basic Spectrogram & Clinical Application: Consonants
Plasticity, exemplars, and the perceptual equivalence of ‘defective’ and non-defective /r/ realisations Rachael-Anne Knight & Mark J. Jones.
Effects of Competence, Exposure, and Linguistic Backgrounds on Accurate Production of English Pure Vowels by Native Japanese and Mandarin Speakers Malcolm.
INTRODUCTION Speech sound disorders of unknown origin (SSDUO) are considered as an output disorder, characterized by incorrect articulation of mostly consonantal.
Phonetic variability of the Greek rhotic sound Mary Baltazani University of Ioannina, Greece  Rhotics exhibit considerable phonetic variety cross-linguistically.
Representing Motion Chapter 2 (pg 30-55). Do Now Why is it important to describe and analyze motion? How fast? How far? Slowing/Speeding? Rest/Constant.
“Speech and the Hearing-Impaired Child: Theory and Practice” Ch. 13 Vowels and Diphthongs –Vowels are formed when sound produced at the glottal source.
Perception of syllable prominence by listeners with and without competence in the tested language Anders Eriksson 1, Esther Grabe 2 & Hartmut Traunmüller.
Niebuhr, D‘Imperio, Gili Fivela, Cangemi 1 Are there “Shapers” and “Aligners” ? Individual differences in signalling pitch accent category.
Prosodic Signalling of (Un)Expected Information in South Swedish Gilbert Ambrazaitis Linguistics and Phonetics Centre for Languages and Literature.
The Human Voice. I. Speech production 1. The vocal organs
Speech Group INRIA Lorraine
Vocal Emotion Recognition with Cochlear Implants Xin Luo, Qian-Jie Fu, John J. Galvin III Presentation By Archie Archibong.
Research on teaching and learning pronunciation
The trough effect: Can we predict tongue lowering from acoustic data alone? Yolanda Vazquez Alvarez.
Stop Place Contrasts before Liquids Edward Flemming MIT.
Chapter 2 Motion in One Dimension. Dynamics The branch of physics involving the motion of an object and the relationship between that motion and other.
Motion in One Dimension
Chapter 2 Motion in One Dimension. Dynamics The branch of physics involving the motion of an object and the relationship between that motion and other.
Motion in One Dimension. The branch of physics involving the motion of an object and the relationship between that motion and other physics concepts Kinematics.
Motion in One Dimension
Present Experiment Introduction Coarticulatory Timing and Lexical Effects on Vowel Nasalization in English: an Aerodynamic Study Jason Bishop University.
Phonetics and Phonology
Abstract Research Questions The present study compared articulatory patterns in production of dental stop [t] with conventional dentures to productions.
Adding Vectors, Rules When two vectors are added, the sum is independent of the order of the addition. This is the Commutative Law of Addition.
Kinematics of Two-Dimensional Motion. Positions, displacements, velocities, and accelerations are all vector quantities in two dimensions. Position Vectors.
Adaptive Design of Speech Sound Systems Randy Diehl In collaboration with Bjőrn Lindblom, Carl Creeger, Lori Holt, and Andrew Lotto.
Speech Science Fall 2009 Oct 28, Outline Acoustical characteristics of Nasal Speech Sounds Stop Consonants Fricatives Affricates.
Connected speech processes Coarticulation Suprasegmentals.
Chapter 1 Introduction to Statistics. Statistical Methods Were developed to serve a purpose Were developed to serve a purpose The purpose for each statistical.
Acoustic Cues to Laryngeal Contrasts in Hindi Susan Jackson and Stephen Winters University of Calgary Acoustics Week in Canada October 14,
♥♥♥♥ 1. Intro. 2. VTS Var.. 3. Method 4. Results 5. Concl. ♠♠ ◄◄ ►► 1/181. Intro.2. VTS Var..3. Method4. Results5. Concl ♠♠◄◄►► IIT Bombay NCC 2011 : 17.
Calibration of Consonant Perception in Room Reverberation K. Ueno (Institute of Industrial Science, Univ. of Tokyo) N. Kopčo and B. G. Shinn-Cunningham.
Epenthetic vowels in Japanese: a perceptual illusion? Emmanual Dupoux, et al (1999) By Carl O’Toole.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT OSMAN BIN SAIF Session 16.
KINEMATICS/KINETICS CORRELATIONS OF ARM MOTOR CONTROL DURING CORIOLIS PERTURBATIONS. A. Pierobon, S.B. Bortolami, J.R. Lackner*, P. DiZio. Ashton Graybiel.
A Study of Taiwanese High School Students' Production and Perception Performance in English Non-High Front Vowels Graduate Student: Wan-chun Tseng Advisor:
11/23/2015Slide 1 Using a combination of tables and plots from SPSS plus spreadsheets from Excel, we will show the linkage between correlation and linear.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Module 6Aberration and Doppler Shift of Light1 Module 6 Aberration and Doppler Shift of Light The term aberration used here means deviation. If a light.
Tongue movement kinematics in speech: Task specific control of movement speed Anders Löfqvist Haskins Laboratories New Haven, CT.
Tone, Accent and Quantity October 19, 2015 Thanks to Chilin Shih for making some of these lecture materials available.
LIN 3201 Sounds of Human Language Sayers -- Week 1 – August 29 & 31.
Section 1Motion Bellringer, continued 1. For each of the devices above, indicate whether it measures distance, time, or speed. 2. For each of the devices.
0 / 27 John-Paul Hosom 1 Alexander Kain Brian O. Bush Towards the Recovery of Targets from Coarticulated Speech for Automatic Speech Recognition Center.
Phonetics: consonants
IIT Bombay 17 th National Conference on Communications, Jan. 2011, Bangalore, India Sp Pr. 1, P3 1/21 Detection of Burst Onset Landmarks in Speech.
Acoustic Phonetics 3/14/00.
Phone-Level Pronunciation Scoring and Assessment for Interactive Language Learning Speech Communication, 2000 Authors: S. M. Witt, S. J. Young Presenter:
Chapter 3 Accelerated Motion. Introduction In this chapter we will examine acceleration and define it in terms of velocity. We will also solve problems.
Introduction to English Pronunciation
Speech Intelligibility and Sentence Duration as a Function of Mode of Communication in Cochlear Implanted Children Nicole L. Wiessner 1, Kristi A. Buckley.
The effect of speech timing on velopharyngeal function
Acoustic to Articoulatory Speech Inversion by Dynamic Time Warping
4aPPa32. How Susceptibility To Noise Varies Across Speech Frequencies
The Human Voice. 1. The vocal organs
In this section you will:
Correlational and Regressive Analysis of the Relationship between Tongue and Lips Motion - An EMA and Video Study of Selected Polish Speech Sounds Robert.
Elaine R. Hitchcocka, Ph.D., Laura L. Koenigb,c, Ph.D.
The Human Voice. 1. The vocal organs
Motion in One Dimension
Jessica McKee Speech, Language and Hearing Sciences
Patricia Keating, Marco Baroni, Sven Mattys, Rebecca Scarborough,
Speech Perception (acoustic cues)
A Japanese trilogy: Segment duration, articulatory kinematics, and interarticulator programming Anders Löfqvist Haskins Laboratories New Haven, CT.
Acceleration 3.1 Changing Velocity
Presentation transcript:

Interarticulator programming in VCV sequences: Effects of closure duration on lip and tongue coordination Anders Löfqvist Haskins Laboratories New Haven, CT

Question In producing a sequence of vowel-labial consonant-vowel, a speaker has to do two or three things: 1. Close and open the lips for the consonant. 2. Move the tongue from the position for the first vowel to that for the second vowel. 3. If the consonant is voiceless, open and close the glottis If the duration of the labial consonant in such a sequence is changed, how is the tongue movement between the two vowels affected? What happens to the interarticulator programming of lip and tongue movements?

Length contrast in japanese When a speaker of Japanese produces the words “kami” and ‘kammi”, the lips are closed for about 60 ms when the consonant is short, and for about 125 ms when the consonant is long.

Background Earlier work on American English has shown that the onset of the tongue movement from the first to the second vowel in such a VCV sequence almost always started before the oral closure for the consonant. In addition, more than 50% of the tongue movement trajectory between the vowels occurred during the oral closure. There was also a weak positive correlation between the magnitude of the tongue movement for the vowels and the interval between tongue movement onset and the onset of the lip closing movement - if you have a long way to go, start early!

Hypotheses If the duration of the oral closure for the consonant is increased, a speaker can in principle use two strategies for the coordination of the tongue and lip movements. 1. Make the same tongue movement trajectory for a long and a short consonant. Such a strategy could have one or two versions, or a combination of them.

Hypotheses, cntd 1a. The onset of the tongue movement might be shifted later relative to the oral closure, so that the tongue movement reached its position for the second vowel at the same point in time for the long and short consonant.

Hypotheses, cntd 1b. Alternatively, the onset of the tongue movement could have the same temporal relationship to the consonant closure. In this case, the interval between the offset of the tongue movement and the release of the oral closure for the labial consonant would thus increase for the long consonant. Another possibility is to shift both the onset and offset of the tongue movement.

Hypotheses, cntd 2. Modify the tongue movement during the long consonant, so that the relative timing between the tongue movement and the lip movements for the oral closure of the consonant would be more or less the same for the long and short consonants. This would imply making a slower tongue movement trajectory for the long than for the short consonant.

Procedure Subjects Five native speakers of Japanese, three male and two female, served as subjects. They reported no speech, language, or hearing problems. They were naive as to the purpose of the study. (The experimental protocol was approved by the IRB at the Yale University School of medicine.)

Procedure, cntd Linguistic material The linguistic material consisted of Japanese words with a sequence of vowel-labial nasal-vowel. These words formed minimal pairs, where the only difference between the pairs was the duration of the labial consonant. The words were designed to require a substantial amount of tongue movement from the first to the second vowel. The following words were used: /kami, kammi/, /kamee, kammee/, /kema, kemma/, /kama, kamma/, /∫ imi, ∫ immi/. The linguistic material was organized into randomized lists and presented to the subjects in Japanese writing, with the words occurring in a short frame sentence. Fifty repetitions of each word were recorded.

Procedure, cntd Movement recording The movements of the lips, the tongue, and jaw were recorded using a three-transmitter magnetometer system; when proper care is taken during the calibration, the spatial resolution of the system is in the order of.5 mm. Receivers were placed on the vermilion border of the upper and lower lip, on three positions of the tongue, referred to as tip, blade and body, and on the lower incisors at the gum line. Two additional receivers placed on the nose and the upper incisors were used for the correction of head movements. All data were subsequently corrected for head movements and rotated to bring the occlusal plane into coincidence with the x axis. This rotation was performed to obtain a uniform coordinate system for all subjects.

Procedure, cntd The articulatory movement signals (induced voltages from the receiver coils) were sampled at 500 Hz after low-pass filtering at 200 Hz. The resolution for all signals was 12 bits A simultaneous acoustic recording was also made with a sampling rate of 10 kHz. Signal processing A measure of lip aperture was obtained by calculating the vertical distance between the upper and lower lip receivers. The velocity and acceleration of this signal were calculated. The onset of the closing movement of the lips for the nasal consonant, was taken as a minimum in the lip aperture acceleration signal just before oral closure.

Procedure, cntd For each tongue receiver, its velocity and speed * were calculated. Tongue movement onsets and offsets were defined algorithmically in the tongue body speed signal as minima during, or close to, the first and second vowels. Measurements - asymmetrical vowel context In this context, landmarks were defined in the acoustic and the tongue movement signals. * Speed:  ( ˙ x 2  ˙ y 2 ) *

The duration of the oral closure for the consonant was measured in the acoustic signal.

Procedure, cntd The magnitude of the tongue movement trajectory from the first to the second vowel was obtained by summing the Euclidean distances between successive samples of the tongue body receiver from movement onset to movement offset. The average speed of the tongue body was obtained by adding the speed of all the individual samples between movement onset and offset and then dividing by the number of samples in the interval.

Oral closure for /m/ “Karewa kami to itta” Subject YK OnsetOffset

Procedure, cntd Three temporal intervals were measured: The interval between tongue movement onset and oral closure; it provides information about the tongue movement relative to consonant closure. The interval between tongue movement onset and lip closing movement onset, defined in the lip aperture acceleration signal; it examines the coordination between tongue and lip movements. The interval between tongue movement offset and oral release; it is useful for examining the tongue movement relative to consonant release.

Oral closure for /m/ “Karewa kami to itta” Subject YK

Procedure, cntd Measurements - symmetrical vowel context In this context, landmarks could not be defined in the tongue movement signals. Instead, the onset and offset of the oral closure, defined in the acoustic signal, were used. The jaw movements were also analyzed, since any tongue movement could be the result of jaw movements associated with the oral closure, in partcular for the open vowel /a/. The duration of the oral closure for the consonant was measured in the acoustic signal. The positions of the jaw and tongue body were measured at the onset and offset of the oral closure. The magnitude of the jaw and tongue body movement paths during closure was measured as defined above.

Results - asymmetrical contexts Oral closure For all speakers, there is a clear and robust difference between the long and short consonants, with the long ones about twice the duration of the short ones.

Closure duration Subject

Results, ctnd Tongue movement duration For most subjects and words, the duration of the tongue body movement from the first to the second vowel is longer in the long than in the short consonant.

Duration of the of the tongue body movement from the first to the second vowel

Results, ctnd Average speed of tongue movement Overall, the average speed of the tongue body movement from the first to the second vowel is slower for the long than for the short consonant.

Average speed of the tongue body movement from the first to the second vowel

Results, ctnd The magnitude of the path of the tongue body movement from the first to the second vowel did not vary systematically with consonant length within or across speakers. Any difference in movement paths was not related to a consistent change in movement onset or offset position.

Path of the tongue body movement from the first to the second vowel.

Results, ctnd Onset of tongue movement relative to oral closure This interval showed no clear pattern for the long and short consonants. Of note is that with only one exception, /kammee/ for subject KN, the tongue movement onset always occurs before the closure.

Interval between onset of tongue movement and oral closure

Results, ctnd Relationship between tongue and lip movement onsets The overall results suggest that the lips lead, but there is no clear difference between the long and short consonants across words and speakers.

Interval between onsets of tongue and lip movements

Results, ctnd Offset of tongue movement relative to oral release The offset of the tongue movement tended to occur after the oral release. The pattern was, however, quite variable as shown by the standard deviations.

Interval between offset of tongue movement and oral release

Results - symmetrical contexts Words with symmetrical vowel contexts were only recorded in three of the subjects. Path of jaw and tongue body movement during oral closure In this context, the path was much shorter than in the asymmetrical context. As expected, the path of both jaw and tongue movement was longer in the context of the open vowel /a/, where it was also longer for the long consonant. The jaw movement was related to the closure for the labial consonant. In the context of the high vowel /i/, the longer movement trajectory of the tongue body for the short consonant was due to coarticulatory influences.

Path of tongue body movement during oral closure Path of jaw movement during oral closure

Summary and conclusions It was hypothesized that a speaker could use one of two possible strategies of interarticulator programming if the duration of the consonant was increased. In one of them, the tongue movement trajectory would be similar for the long and short consonants; thus, the timing between the lip and tongue movements would change. In the second one, the tongue movement would be modified for the long consonant, thus resulting in a similar coordination of lip and tongue movements for both the long and short consonants. The present results clearly support the second strategy.

They thus show that speakers of Japanese modify their tongue movements when they produce a sequence of a vowel-bilabial consonant-vowel with a long and a short consonant. In particular, the duration of the tongue movement is increased, while its average speed is reduced in the long consonant. However, the path of the tongue movement didn’t vary consistently with consonant duration. Thus, the coordination of the tongue movement relative to the onset and release of the oral closure was similar for the long and short consonants. Importantly, all five speakers showed the same modification of tongue movement speed and duration for the long consonant. These finding are similar to the ones reported by Caroline Smith in her Yale 1992 dissertation. They thus suggest that the Japanese speakers tend to maintain a similar, but not identical, coordination between the tongue and lip movements across consonant length

The present results also show that speakers of Japanese have the same coordination between lip and tongue movements in VCV sequences as speakers of American English, for both the long and short consonants in Japanese. At the same time, American English and Japanese are described to have stress- and mora-timing, respectively, but the interarticulator programming of lip and tongue movements appears to be virtually identical for the two languages. This raises the issue of whether the prosodic structure of a language will influence patterns of interarticulator timing. The results presented by Caroline Smith suggest that this may be the case, since speakers of Japanese and Italian, which also has a length contrast, may have different patterns of coordination in VCV sequences when the consonant length differs. Japanese is described as having mora timing, Italian is said to have syllable timing. The obvious next step, examining Italian speakers using the same experimental approach, is currently under way.

Acknowledgments I am grateful to Mariko Yanagawa for help with the Japanese material and running the experiments. This work was supported by Grant No. DC from the National Institute on Deafness and Other Communication Disorders, National Institutes of Health.