SPEECH PERCEPTION DAY 16 – OCT 2, 2013 Brain & Language LING NSCI Harry Howard Tulane University
Course organization The syllabus, these slides and my recordings are available at If you want to learn more about EEG and neurolinguistics, you are welcome to participate in my lab. This is also a good way to get started on an honor's thesis. The grades are posted to Blackboard. 10/02/13Brain & Language, Harry Howard, Tulane University 2
REVIEW Speech recognition 10/02/13Brain & Language, Harry Howard, Tulane University 3
Linguistic model, Fig. 2.1 p /02/13Brain & Language, Harry Howard, Tulane University 4 Discourse model Syntax Sentence prosody Morphology Word prosody Segmental phonology perception Segmental phonology perception Acoustic phonetics Feature extraction Segmental phonology production Segmental phonology production Articulatory phonetics Speech motor control INPUT Sentence level Word level
Summary of speech recognition Details in Ingram It is bottom up. Lexical knowledge of the input would be helpful, but it doesn’t make the process much easier. “Speech perception involves phonological parsing prior to lexical access.” Its output are phonological items/targets. These targets are abstract They include both acoustic and articulatory information [features]. They are normalized across rate, style, and speaker. They may be underspecified. They are discrete. … and organized into a hierarchy. It is/may be different from other kinds of auditory perception … … so it may require special, species-specific neural machinery. 10/02/13Brain & Language, Harry Howard, Tulane University 5
SPEECH PERCEPTION Ingram §6 10/02/13Brain & Language, Harry Howard, Tulane University 6
The speech mode hypothesis (SMH) Speed of speech vs. non-speech perception The articulators move at the rate of 8-10 phones per second = 3-4 syllables per second = 2-3 words per second. Early researchers tried to build a machine for reading to the blind [the auditory cipher] which converted each letter to a different non- speech sound and found that no one could understand it at greater than 2-3 sounds per second, which is about a quarter of the rate at which speech is understood. What to do about this inconsistency? 10/02/13Brain & Language, Harry Howard, Tulane University 7
Speech is different! How so? Speech perception is different from other forms of auditory perception because its targets are linked to a specialized system for their production … … which we might engage when we listen to speech. Motor theory of speech perception Speech mode hypothesis 10/02/13Brain & Language, Harry Howard, Tulane University 8
Can you turn your speech mode off? What is this?It is this. 10/02/13Brain & Language, Harry Howard, Tulane University 9
Weak vs. strong versions of SMH What does such a distinction mean? The weak speech mode hypothesis When we listen to speech, we engage our knowledge of language. The strong speech mode hypothesis When we listen to speech, we engage perceptual mechanisms specialized for speech. In the next few slides, we go over the evidence for the latter. 10/02/13Brain & Language, Harry Howard, Tulane University 10
Dichotic listening Tendencies of right-ear advantage by speech sound No advantage Weak right-ear advantage Strong right-ear advantage vowels liquids (l,r), glides (j,w), fricatives stops the acoustic cues for vowels do not depend on context the acoustic cues for consonants depend on context [see p. 116] > special machinery? 10/02/13Brain & Language, Harry Howard, Tulane University 11
Intermission: Voice onset time (VOT) 10/02/13Brain & Language, Harry Howard, Tulane University 12 [b] [p] [p ʰ ]
Categorical perception 10/02/13Brain & Language, Harry Howard, Tulane University 13 Chinchillas do this too! The Clinton-Kennedy continuum
NEXT TIME Finish Ingram §6. ☞ Go over questions at end of chapter. 10/02/13Brain & Language, Harry Howard, Tulane University 14