Motor Theory + Signal Detection Theory

Slides:



Advertisements
Similar presentations
Reading and the phonetic module Carol A. Fowler Haskins Laboratories University of Connecticut Yale University.
Advertisements

Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
CS 551/651: Structure of Spoken Language Lecture 12: Tests of Human Speech Perception John-Paul Hosom Fall 2008.
Copyright © 2009 Pearson Education Canada6-1 Chapter 6: Cognition in Infants and Toddlers 6.1 Piaget’s Theory 6.2 Information Processing 6.3 Language MODULES.
Chapter 12 Speech Perception. Animals use sound to communicate in many ways Bird calls Bird calls Whale calls Whale calls Baboons shrieks Baboons shrieks.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Language Comprehension Speech Perception Semantic Processing & Naming Deficits.
Cognition, 8e by Margaret W. MatlinChapter 2 Cognition, 8e Chapter 2 Perceptual Processes I: Visual and Auditory Recognition.
SPEECH PERCEPTION 2 DAY 17 – OCT 4, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
Ling 240: Language and Mind Acquisition of Phonology.
Speech perception 2 Perceptual organization of speech.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
AS14.1 – The Role of Neural Oscillations in Auditory Resolution and Speech Perception – Implications for CAPD, Phonological Awareness and Reading Disorders.
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Profile of Phoneme Auditory Perception Ability in Children with Hearing Impairment and Phonological Disorders By Manal Mohamed El-Banna (MD) Unit of Phoniatrics,
Neural correlates of continuous and categorical sinewave speech perception: An FMRI study Rutvik Desai, Einat Liebenthal, Eric Waldron, and Jeffrey R.
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Language. Using Language What is language for? Using Language What is language for? – Rapid, efficient communication To accomplish this goal, what needs.
Organizational Notes no study guide no review session not sufficient to just read book and glance at lecture material midterm/final is considered hard.
PSY 369: Psycholinguistics
SPEECH PERCEPTION The Speech Stimulus Perceiving Phonemes Top-Down Processing Is Speech Special?
Speech Perception Richard Wright Linguistics 453.
What is Cognitive Science? … is the interdisciplinary study of mind and intelligence, embracing philosophy, psychology, artificial intelligence, neuroscience,
What is Cognitive Science? … is the interdisciplinary study of mind and intelligence, embracing philosophy, psychology, artificial intelligence, neuroscience,
Language Comprehension Speech Perception Naming Deficits.
Perceptual Inference and Information Integration in Brain and Behavior PDP Class Jan 11, 2010.
The Perception of Speech
Language Comprehension Speech Perception Meaning Representation.
Conclusions  Constriction Type does influence AV speech perception when it is visibly distinct Constriction is more effective than Articulator in this.
Chapter 13: Speech Perception
Speech Perception. Phoneme - a basic unit of a speech sound that distinguishes one word from another Phonemes do not have meaning on their own but they.
Speech Perception 4/6/00 Acoustic-Perceptual Invariance in Speech Perceptual Constancy or Perceptual Invariance: –Perpetual constancy is necessary, however,
The Motor Theory of Speech Perception April 1, 2013.
Speech Or can you hear me now?. Linguistic Parts of Speech Phone Phone Basic unit of speech sound Basic unit of speech sound Phoneme Phoneme Phone to.
Mr Background Noise and Miss Speech Perception in: by Elvira Perez and Georg Meyer.
Motor Theory Remnants April 3, 2012 Dirty Work Project Reports #5 to turn in. On Thursday, we’ll talk about the muscles that control articulation… And.
Perceiving Transformational Invariance Discovering the basis multi-sensory integration Richard Hurtig Department of Speech pathology & Audiology The University.
Speech Perception 4/4/00.
Glides, Place and Perception March 18, 2010 News The hard drive on the computer has been fixed! A couple of new readings have been posted to the course.
More Motor Theory + Fricative Acoustics March 30, 2010.
Transitions + Perception March 27, 2012 Tidbits First: Guidelines for the final project report So far, I have two people who want to present their projects.
Exemplar Theory April 10, 2014 Practicalities Project presentations/review session: Thursday, April 17 th, 1-2:30 pm Science A 147 Final exam: Thursday,
Perception + Vocal Tract Physiology November 25, 2014.
Epenthetic vowels in Japanese: a perceptual illusion? Emmanual Dupoux, et al (1999) By Carl O’Toole.
Pragmatically-guided perceptual learning Tanya Kraljic, Arty Samuel, Susan Brennan Adaptation Project mini-Conference, May 7, 2007.
Chapter 8: Perceiving Motion
Information Processing Assumptions Measuring the real-time stages General theory –structures –control processes Representation –definition –content vs.
Sensation & Perception
How Does auditory perception organization works ? by Elvira Perez and Georg Meyer Dept. Psychology, Liverpool University, UK Hoarse Meeting, Chrysler Ulm,
Phonetic Context Effects Major Theories of Speech Perception Motor Theory: Specialized module (later version) represents speech sounds in terms of intended.
Hearing Research Center
Acoustic Continua and Phonetic Categories Frequency - Tones.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
CSD 2230 INTRODUCTION TO HUMAN COMMUNICATION DISORDERS Normal Sound Perception, Speech Perception, and Auditory Characteristics at the Boundaries of the.
Chapter 13: Speech Perception. The Acoustic Signal Produced by air that is pushed up from the lungs through the vocal cords and into the vocal tract Vowels.
Basic Cognitive Processes - 2
Exemplar Theory, part 2 April 15, 2013.
Language Perception.
WebCT You will find a link to WebCT under the “Current Students” heading on It is your responsibility to know how to work WebCT!
Transitions + Perception March 25, 2010 Tidbits Mystery spectrogram #3 is now up and ready for review! Final project ideas.
AUDITORY CORTEX 1 SEPT 11, 2015 – DAY 8 Brain & Language LING NSCI Fall 2015.
Motor Theory of Perception March 29, 2012 Tidbits First: Guidelines for the final project report So far, I have two people who want to present their.
Signal Detection Theory March 25, 2010 Phonetics Fun, Ltd. Check it out:
Against formal phonology (Port and Leary).  Generative phonology assumes:  Units (phones) are discrete (not continuous, not variable)  Phonetic space.
Solve this maze at your leisure. Start at phil’s house. At first, you can only make right turns through the maze. Each time you cross the red zigzag sign.
What can we expect of cochlear implants for listening to speech in noisy environments? Andrew Faulkner: UCL Speech Hearing and Phonetic Sciences.
Danielle Werle Undergraduate Thesis Intelligibility and the Carrier Phrase Effect in Sinewave Speech.
Perception and Attention Advanced Cognitive Psychology PSY 421, Fall 2004.
Speech Perception (acoustic cues)
Presentation transcript:

Motor Theory + Signal Detection Theory March 23, 2010

Oh Yeahs. Nasometer labs due! Dental vs. alveolar vs. bilabial release bursts. Examples from Yanyuwa: Examples from Hindi: Creating synthetic formant transitions: KlattTalk. Yanyuwa: alveolar, bilabial, dental, retroflex Hindi: bilabial, dental, retroflex (voiced on top, voiceless on bottom)

KlattTalk KlattTalk has since become the standard for formant synthesis. (DECTalk) http://www.asel.udel.edu/speech/tutorials/synthesis/vowels.html

Categorical Perception continuous physical distinctions are perceived in discrete categories. In the in-class perception experiment: There were 11 different syllable stimuli They only differed in the locus of their F2 transition F2 Locus range = 726 - 2217 Hz Source: http://www.ling.gu.se/~anders/KatPer/Applet/index.eng.html

Example stimuli from the in-class experiment. Stimulus #1 Stimulus #6 Stimulus #11 Example stimuli from the in-class experiment.

Identification In Categorical Perception: All stimuli within a category boundary should be labeled the same.

Discrimination Original task: ABX discrimination Stimuli across category boundaries should be 100% discriminable. Stimuli within category boundaries should not be discriminable at all. In practice, categorical perception means: the discrimination function can be determined from the identification function.

Identification  Discrimination Let’s consider a case where the two sounds in a discrimination pair are the same. Example: the pair is stimulus 3 followed by stimulus 3 Identification data--Stimulus 3 is identified as: [b] 95% of the time [d] 5% of the time The discrimination pair will be perceived as: [b] - [b] - .95 * .95 = .9025 [d] - [d] - .05 * .05 = .0025 Probability of same response is predicted to be: (.9025 + .0025) = .905 = 90.5%

Identification  Discrimination Let’s consider a case where the two sounds in a discrimination pair are different. Example: the pair is stimulus 9 followed by stimulus 11 Identification data: Stimulus 9: [d] 80% of the time, [g] 20% of the time Stimulus 11: [d] 5% of the time, [g] 95% of the time The discrimination pair will be perceived as: [d] - [d] - .80 * .05 = .04 [g] - [g] - .20 * .95 = .19 Probability of same response is predicted to be: (.04 + .19) = 23%

Discrimination In this discrimination graph-- Solid line is the observed data Dashed line is the predicted data (on the basis of the identification scores) Note: the actual listeners did a little bit better than the predictions.

Categorical, Continued Categorical Perception was also found for stop/glide/vowel distinctions: 10 ms transitions: [b] percept 60 ms transitions: [w] percept 200 ms transitions: [u] percept

Interpretation Main idea: in categorical perception, the mind translates an acoustic stimulus into a phonemic label. (category) The acoustic details of the stimulus are discarded in favor of an abstract representation. A continuous acoustic signal: Is thus transformed into a series of linguistic units:

The Next Level Interestingly, categorical perception is not found for non-speech stimuli. Miyawaki et al: tested perception of an F3 continuum between /r/ and /l/.

The Next Level They also tested perception of the F3 transitions in isolation. Listeners did not perceive these transitions categorically.

The Implications Interpretation: we do not perceive speech in the same way we perceive other sounds. “Speech is special”… and the perception of speech is modular. A module is a special processor in our minds/brains devoted to interpreting a particular kind of environmental stimuli.

Module Characteristics You can think of a module as a “mental reflex”. A module of the mind is defined as having the following characteristics: Domain-specific Automatic Fast Hard-wired in brain Limited top-down access (you can’t “unperceive”) Example: the sense of vision operates modularly.

A Modular Mind Model central processes judgment, imagination, memory, attention vision hearing touch speech modules transducers eyes ears skin etc. external, physical reality

Remember this stuff? Speech is a “special” kind of sound because it exhibits spectral change over time.  it’s processed by the speech module, not by the auditory module.

SWS Findings The uninitiated either hear sinewave speech as speech or as “whistles”, “chirps”, etc. Claim: once you hear it as speech, you can’t go back. The speech module takes precedence (Limited top-down access) Analogy: it’s impossible to not perceive real speech as speech. We can’t hear the individual formants as whistles, chirps, etc. Motor theory says: we don’t perceive the “sounds”, we perceive the gestures which shape the spectrum.

McGurk Videos

McGurk Effect explained Audio Visual Perceived ba + ga  da ga + ba  ba (bga) Some interesting facts: The McGurk Effect is exceedingly robust. Adults show the McGurk Effect more than children. Americans show the McGurk Effect more than Japanese.

Original McGurk Data Auditory Visual Stimulus: ba-ba ga-ga Response types: Auditory: ba-ba Fused: da-da Visual: ga-ga Combo: gabga, bagba Age Auditory Visual Fused Combo 3-5 19% 36 81 0 7-8 36 0 64 0 18-40 2 0 98 0

Original McGurk Data Auditory Visual Stimulus: ga-ga ba-ba Response types: Auditory: ba-ba Fused: da-da Visual: ga-ga Combo: gabga, bagba Age Auditory Visual Fused Combo 3-5 57% 10 0 19 7-8 36 21 11 32 18-40 11 31 0 54

Audio-Visual Sidebar Visual cues affect the perception of speech in non-mismatched conditions, as well. Scientific studies of lipreading date back to the early twentieth century The original goal: improve the speech perception skills of the hearing-impaired Note: visual speech cues often complement audio speech cues In particular: place of articulation However, training people to become better lipreaders has proven difficult… Some people got it; some people don’t.

Sumby & Pollack (1954) First investigated the influence of visual information on the perception of speech by normal-hearing listeners. Method: Presented individual word tokens to listeners in noise, with simultaneous visual cues. Task: identify spoken word Clear: +10 dB SNR: + 5 dB SNR: 0 dB SNR:

Sumby & Pollack data Auditory-Only Audio-Visual Visual cues provide an intelligibility boost equivalent to a 12 dB increase in signal-to-noise ratio.

Tadoma Method Some deaf-blind people learn to perceive speech through the tactile modality, by using the Tadoma method.

Audio-Tactile Perception Fowler & Dekle: tested ability of (naive) college students to perceive speech through the Tadoma method. Presented synthetic stops auditorily Combined with mismatched tactile information: Ex: audio /ga/ + tactile /ba/ Also combined with mismatched orthographic information: Ex: audio /ga/ + orthographic /ba/ Task: listeners reported what they “heard” Tactile condition biased listeners more towards “ba” responses

Fowler & Dekle data orthographic mismatch condition read “ba” tactile mismatch condition read “ba” felt “ba”

fMRI data Benson et al. (2001) Non-Speech stimuli = notes, chords, and chord progressions on a piano

fMRI data Benson et al. (2001) Difference in activation for natural speech stimuli versus activiation for sinewave speech stimuli

Mirror Neurons In the 1990s, researchers in Italy discovered what they called “mirror neurons” in the brains of macaques. Macaques had been trained to make grasping motions with their hands. Researchers recorded the activity of single neurons while the monkeys were making these motions. Serendipity: the same neurons fired when the monkeys saw the researchers making grasping motions.  a neurological link between perception and action. Motor theory claim: same links exist in the human brain, for the perception of speech gestures

Motor Theory, in a nutshell The big idea: We perceive speech as abstract “gestures”, not sounds. Evidence: The perceptual interpretation of speech differs radically from the acoustic organization of speech sounds Speech perception is multi-modal Direct (visual, tactile) information about gestures can influence/override indirect (acoustic) speech cues Limited top-down access to the primary, acoustic elements of speech