Chapter 12 Speech Perception. Animals use sound to communicate in many ways Bird calls Bird calls Whale calls Whale calls Baboons shrieks Baboons shrieks.

Slides:



Advertisements
Similar presentations
Phonetics as a scientific study of speech
Advertisements

Chapter 12: Speech and Music Perception
Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
Jennifer Nazar.  A language disorder produced by brain damage.  Most studies come from those who have aphasias.  Study behaviors associated with the.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Language Comprehension Speech Perception Semantic Processing & Naming Deficits.
SPEECH PERCEPTION 2 DAY 17 – OCT 4, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
Speech and Language. It is the highest function of the nervous system Involves understanding of spoken & printed words It is the ability to express ideas.
The Neuroscience of Language. What is language? What is it for? Rapid efficient communication – (as such, other kinds of communication might be called.
Speech perception 2 Perceptual organization of speech.
The Cerebral Cortex. The Evolving Brain Different animal species have many structures in common, including a cerebellum and cortex. The cortex is much.
1. Why was Ms. Jensen worried about her son? 2. What is significant about teen’s frontal lobes? 3. What is the role of the frontal lobe? 4. What are some.
Development of Speech Perception. Issues in the development of speech perception Are the mechanisms peculiar to speech perception evident in young infants?
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Chapter 6 Perception.
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Language within our grasp: Gesture, mirror neurons, and meaning.
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Language. Using Language What is language for? Using Language What is language for? – Rapid, efficient communication To accomplish this goal, what needs.
Cognitive Processes PSY 334 Chapter 2 – Perception April 9, 2003.
Anatomy of language processing Mark Harju. Most components for language processing are located in the left hemisphere Most components for language processing.
1 Pattern Recognition (cont.). 2 Auditory pattern recognition Stimuli for audition is alternating patterns of high and low air pressure called sound waves.
PSY 369: Psycholinguistics
Development of Language and Symbols
SPEECH PERCEPTION The Speech Stimulus Perceiving Phonemes Top-Down Processing Is Speech Special?
Language Comprehension Speech Perception Naming Deficits.
Auditory-acoustic relations and effects on language inventory Carrie Niziolek [carrien] may 2004.
The Perception of Speech
Communication Disorders
Psycholinguistics.
Chapter 13: Speech Perception
COMMUNICATING WITH OTHERS The Development and Use of Language
Phonetics: the generation of speech Phonemes “The shortest segment of speech that, if changed, would change the meaning of a word.” hog fog log *Phonemes.
Chapter 6 Perception.
1 Speech Perception 3/30/00. 2 Speech Perception How do we perceive speech? –Multifaceted process –Not fully understood –Models & theories attempt to.
 The origin of grammatical rules is ascribed to an innate system in the human brain.  The knowledge of and competence for human language is acquired.
Lipreading: how it works. Learning objectives Recognise the different processes and skills involved in lipreading Revise factors that help or hinder lipreading.
Speech Perception 4/4/00.
LATERALIZATION OF PHONOLOGY 2 DAY 23 – OCT 21, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
Sensation & Perception
Cognition 7e, Margaret MatlinChapter 13 The Development of Language Language in Infants rate of acquisition – approx 7 words/day, birth-6 vocabulary size.
Phonetic Context Effects Major Theories of Speech Perception Motor Theory: Specialized module (later version) represents speech sounds in terms of intended.
1.  What is Speech ?  Speech is complex form of communication in which spoken words convey ideas.  When we speak, first we understand. 2.
The Holes in the Brian Help Us Sort Out Sounds..  I. The Brain’s ability to sort out sounds  1. speech sounds are categorized.  2.Misinterpretations.
Sounds and speech perception Productivity of language Speech sounds Speech perception Integration of information.
CSD 2230 INTRODUCTION TO HUMAN COMMUNICATION DISORDERS Normal Sound Perception, Speech Perception, and Auditory Characteristics at the Boundaries of the.
PHYSIOLOGY OF SPEECH Prof. Sultan Ayoub Meo
Chapter 13: Speech Perception. The Acoustic Signal Produced by air that is pushed up from the lungs through the vocal cords and into the vocal tract Vowels.
Speech Perception.
Language Perception.
WebCT You will find a link to WebCT under the “Current Students” heading on It is your responsibility to know how to work WebCT!
The Cerebral Cortex. The Evolving Brain Different animal species have many structures in common, including a cerebellum and cortex. The cortex is much.
Language and the brain Introduction to Linguistics.
Brain and Language. Separating the two hemispheres Commisurotomy: sever the corpus callosum Left visual field disconnected from language center ● CLAP.
The two different parts of speech Speech Production Speech Perception.
Chapter 11 Language. Some Questions to Consider How do we understand individual words, and how are words combined to create sentences? How can we understand.
Solve this maze at your leisure. Start at phil’s house. At first, you can only make right turns through the maze. Each time you cross the red zigzag sign.
Branches of Linguistics
Speechreading Based on Tye-Murray (1998) pp
During the first 6 months of life, a baby’s vocalisations are dominated by crying, cooing, and laughter, which have limited value as a mean of communication.
Physiology of Language
The Cerebral Cortex.
THE HUMAN BRAIN.
Language: An Overview Language is a brain function
Prefrontal Cortex “Executive functions”: planning, behavioral inhibition, emotional regulation, forward thinking, achieving goals, personality expression,
The Cerebral Cortex Thin outer layer of interconnected neurons that are responsible for higher level thinking & skills.
Speech Perception.
NeuroLinguistics Mam Lubna Umar.
UNIT 3 THE CONSCIOUS SELF
Multisensory integration: perceptual grouping by eye and ear
Presentation transcript:

Chapter 12 Speech Perception

Animals use sound to communicate in many ways Bird calls Bird calls Whale calls Whale calls Baboons shrieks Baboons shrieks Vervet calls Vervet calls Grasshopper rubbing legs Grasshopper rubbing legs These kinds of communication differ from language in the structure of the signals. These kinds of communication differ from language in the structure of the signals.

Speech perception is a broad category Understanding what is said (linguistic information) Understanding what is said (linguistic information) Understanding “paralinguistic information” Understanding “paralinguistic information” Speaker’s identity Speaker’s identity Speaker’s affective state Speaker’s affective state Speech processing ≠ linguistics processing. Speech processing ≠ linguistics processing.

Vocal tract Includes larynx, throat, tongue, teeth, and lips. Includes larynx, throat, tongue, teeth, and lips. Vocal chords = vocal folds Vocal chords = vocal folds Male vocal chords 60% larger than female vocal chords in humans Male vocal chords 60% larger than female vocal chords in humans Size of vocal chords are not the sole cue to sex of speaker. Children’s voices can be discriminated. Size of vocal chords are not the sole cue to sex of speaker. Children’s voices can be discriminated.

Physical disturbances in air ≠ phonemes Many different sounds are lumped together in a every single phoneme. Many different sounds are lumped together in a every single phoneme. Another case of separating the physical from the psychological. Another case of separating the physical from the psychological.

Humans normally speak at about 12 phonemes per second. Humans normally speak at about 12 phonemes per second. Humans can comprehend speech at up to about 50 phonemes per second. Humans can comprehend speech at up to about 50 phonemes per second. Voice spectrogram changes with age. Voice spectrogram changes with age. Spectrograms can be taken of all sorts of sounds. Spectrograms can be taken of all sorts of sounds.

Neural analysis of speech sounds One phoneme can have distinct sound spectrograms. Distinct sound spectrograms can be metamers for a phoneme. One phoneme can have distinct sound spectrograms. Distinct sound spectrograms can be metamers for a phoneme.

Brain mechanisms of speech perception

Brain mechanisms of speech perception Single-cell recordings in monkeys show they are sensitive to: 1. Time lapsing between lip movements and start of sound production 2. Acoustic context of sound 3. Rate of sound frequency changes

Human studies Human studies have been based on neuroimaging (fMRI and PET). Human studies have been based on neuroimaging (fMRI and PET). A1 is not a linguistic center; merely an auditory center. It does not respond preferentially to speech, rather than sound. A1 is not a linguistic center; merely an auditory center. It does not respond preferentially to speech, rather than sound. Speech processing is a grab bag of kinds of processing, e.g. linguistic, emotional, and speaker identity. Speech processing is a grab bag of kinds of processing, e.g. linguistic, emotional, and speaker identity.

Wernicke’s aphasia Subjects can hear sounds. Subjects can hear sounds. Subjects lose ability to comprehend speech, though they can produce (clearly disturbed) speech themselves. Subjects lose ability to comprehend speech, though they can produce (clearly disturbed) speech themselves.

Other brain regions involved in speech processing Right temporal hemisphere is involved in emotion, speaker sex, and identity. Right temporal hemisphere is involved in emotion, speaker sex, and identity. Phonagnosia Phonagnosia Right temporal hemisphere is less involved in linguistic analysis. Right temporal hemisphere is less involved in linguistic analysis. Right pre-frontal cortex and parts of the limbic systems respond to emotion. Right pre-frontal cortex and parts of the limbic systems respond to emotion.

Other brain regions involved in speech processing Both hemispheres active in human vocalizations, such as laughing or humming. Both hemispheres active in human vocalizations, such as laughing or humming. Some motor areas for speech are active during speech perception. Some motor areas for speech are active during speech perception.

A “what” and “where” pathway in speech processing? One pathway is anterior (forward) and ventral (below) One pathway is anterior (forward) and ventral (below) The other pathway is posterior (backward) and dorsal (above). The other pathway is posterior (backward) and dorsal (above). Not clear what these pathways do. Not clear what these pathways do.

Understanding speech: Aftereffects Tilt aftereffect and motion aftereffect due to “fatigue” of specific neurons. Tilt aftereffect and motion aftereffect due to “fatigue” of specific neurons. Eimas & Corbett, (1973), performed a linguistic version. Eimas & Corbett, (1973), performed a linguistic version. Take ambiguous phonemes, e.g. between /t/ and /d/. Take ambiguous phonemes, e.g. between /t/ and /d/. Listen to /d/ over and over, then the ambiguity disappears. Listen to /d/ over and over, then the ambiguity disappears.

Understanding speech: Context effects In vision, surrounding objects affect interpretation of size, color, brightness. In other words, context influences perception. In vision, surrounding objects affect interpretation of size, color, brightness. In other words, context influences perception. In speech, context influences perception. We noted this earlier with /di/ and /du/. In speech, context influences perception. We noted this earlier with /di/ and /du/.

Understanding speech: Context effects Semantic context can influence perception. Semantic context can influence perception. Examples of song lyrics. Examples of song lyrics. Speed of utterance influences phonetic interpretation. Speed of utterance influences phonetic interpretation. A syllable may sound like /ba/ when preceding words are spoken slowly, but like /pa/ when preceding words are spoken quickly. A syllable may sound like /ba/ when preceding words are spoken slowly, but like /pa/ when preceding words are spoken quickly. Cadence of a sentence can influence interpretation of the last word. (Ladeford & Broadbent, 1957) Cadence of a sentence can influence interpretation of the last word. (Ladeford & Broadbent, 1957)

Understanding speech: visual effects McGurk Effect Movies of speakers influence syllables heard. Movies of speakers influence syllables heard. Vocal /ga/ + lip /ba/ = /da/ Vocal /ga/ + lip /ba/ = /da/ Vocal “tought” + lip “hole” = “towel”. Vocal “tought” + lip “hole” = “towel”. McGurk effect reduced with face inversion McGurk effect reduced with face inversion

Emotions of talking heads Movie of facial emotion + voice with an emotion Movie of facial emotion + voice with an emotion When face and voice agree, most subject correctly identity emotion. When face and voice agree, most subject correctly identity emotion. When face and voice conflict, facial expression provided the emotion. When face and voice conflict, facial expression provided the emotion.

McGurk effect + talking heads effect makes sense, since it enables humans to function more reliably in noise environments. McGurk effect + talking heads effect makes sense, since it enables humans to function more reliably in noise environments. Infants weeks old can match voice and face. Infants weeks old can match voice and face. Humans can match movies of speakers with voices of speakers. Humans can match movies of speakers with voices of speakers.

Monkeys and preferential looking Ghazanfar & Logothetis, (2003). Ghazanfar & Logothetis, (2003). Showed monkeys two silent movies of monkeys vocalizing at the same time. Showed monkeys two silent movies of monkeys vocalizing at the same time. Played a vocalization that matched one of the silent movies. Played a vocalization that matched one of the silent movies. All 20 monkeys looked at the monkey face that matched the sound. All 20 monkeys looked at the monkey face that matched the sound.

More neuroimaging of speech perception Subjects watched faces of silent speakers. Subjects watched faces of silent speakers. MT (aka V5) was active for motion processing. MT (aka V5) was active for motion processing. A1 and additional language centers were also active. A1 and additional language centers were also active.

Perceived sound boundaries in words are illusory. Perceived sound boundaries in words are illusory. “Mondegreens” “Mondegreens” Pauses indicate times at which to switch speakers. Pauses indicate times at which to switch speakers. Disfluency: repetitions, false starts, and useless interjections. Disfluency: repetitions, false starts, and useless interjections. Help by parsing sentence, give subject time to process, and hinting at new information. Help by parsing sentence, give subject time to process, and hinting at new information.

Language-based learning impairment: A specifically linguistic, rather than acoustic impairment. Language-based learning impairment: A specifically linguistic, rather than acoustic impairment.

Fun illusion (nothing to do with class): e.html e.html