Visual speech speeds up the neural processing of auditory speech van Wassenhove, V., Grant, K. W., & Poeppel, D. (2005) Proceedings of the National Academy.

Slides:



Advertisements
Similar presentations
Frequency Band-Importance Functions for Auditory and Auditory- Visual Speech Recognition Ken W. Grant Walter Reed Army Medical Center Washington, D.C.
Advertisements

Cross-modal perception of motion- based visual-haptic stimuli Ian Oakley & Sile OModhrain Palpable Machines Research Group
Helen Gaeta, David Friedman, & Gregory Hunt Cognitive Electrophysiology Laboratory New York State Psychiatric Institute Differential Effects of Stimulus.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
ERP Characteristics of Early AD
Electroencephalogram (EEG) and Event Related Potentials (ERP) Lucy J. Troup 28 th January 2008 CSU Symposium on Imaging.
Early auditory novelty processing in humans: auditory brainstem and middle-latency responses Slabu L, Grimm S, Costa-Faidella J, Escera C.
P3 target cue target long CTI cue target cue short CTI children old PZ cue target Cue-related ERPs.
Infant sensitivity to distributional information can affect phonetic discrimination Jessica Maye, Janet F. Werker, LouAnn Gerken A brief article from Cognition.
Recognition memory amongst individuals varying in the personality dimensions of Reward Seeking and Impulsivity Chase Kluemper 1, Chelsea Black 1, Yang.
Effects of Sound and Visual Congruency on Product Selection and Preference Brock Bass Felipe Fernandez Drew Link Andrew Schmitz.
Charles Spence Department of Experimental Psychology, Oxford University New Perspectives from the Human Sciences Consumer Focus Workshop (November, 2001)
Cross-modal Prediction in Speech Perception Carolina Sánchez, Agnès Alsius, James T. Enns & Salvador Soto-Faraco Multisensory Research Group Universitat.
Experiment 2: MEG Study Materials and Methods: 11 right-handed subjects with 20:20 vision were run. 3 subjects’ data was discarded because of poor performance.
Links Between Action Perception and Action Production in 10-Week-Old Infants Vincent M. Reid 1 and Katharina Kaduk 1 Department of Psychology, Durham University,
INTRODUCTION Emotional stimuli that are irrelevant to task performance may redirect attentional resources, particularly among individuals evidencing high.
Electroencephalography Electrical potential is usually measured at many sites on the head surface.
Contact Introduction Viewing a talking person’s face and mouth may enhance speech understanding.
Sensation Perception = gathering information from the environment 2 stages: –Sensation = simple sensory experiences and translating physical energy from.
Interactions between Stimulus- Specific Adaptation and Visual Auditory Integration in the Forebrain of the Barn Owl Amit Reches, Shai Netser, and Yoram.
Multisensory convergence Traditionally, represented by inputs from more than one sensory modality, each of which can independently activate the target.
Abstract Cognitive control processes reduce the effects of irrelevant or misleading information on performance. We report a study suggesting that effective.
Learning Styles and Comprehension Brought to you by: Jennifer, Annika, and Katharine.
The Time Course of Processing Emotional Prosody: Behavioral and Electrophysiological Investigations Lauren Cornew, 1 Leslie J. Carver, 1 and Tracy Love.
Audiovisual Multisensory Facilitation: A Fresh Look at Neural Coactivation and Inverse Effectiveness. Lynnette Leone North Dakota State University.
Conclusions  Constriction Type does influence AV speech perception when it is visibly distinct Constriction is more effective than Articulator in this.
All slides © S. J. Luck, except as indicated in the notes sections of individual slides Slides may be used for nonprofit educational purposes if this copyright.
Chapter 4 Opener. Figure 4.1 A testing booth set up for the head-turn preference paradigm.
Chapter 4 Opener. Figure 4.1 A testing booth set up for the head-turn preference paradigm.
1 Web Tour. 2 Materials License Web Tour 3 Materials License Moderator View Follow Me Publish URL to Chat Go to URL…
Participants: 57 children (6-8 years old, 35 boys) participated in experiments. All were schoolchildren in first class of elementary school in Novosibirsk,
Visual-auditory interactions assessed with ABR Dr. W. David Hairston, Visual & Auditory Processes Branch Human Research and Engineering Directorate Army.
INTRODUCTION AIMS AND PREDICTIONS METHODS Participants: 18 children (9-10; M = 10). 38 young adults (20-30; M = 24) 26 older adults (65-85; M = 72) EEG.
Infant Discrimination of Voices: Predictions from the Intersensory Redundancy Hypothesis Lorraine E. Bahrick, Robert Lickliter, Melissa A. Shuman, Laura.
As expected, a large N400 effect was observed for all 3 word types in both experiments, |ts|≥7.69, ps
Reicher (1969): Word Superiority Effect Dr. Timothy Bender Psychology Department Missouri State University Springfield, MO
N400-like semantic incongruity effect in 19-month-olds: Processing known words in picture contexts Manuela Friedrich and Angela D. Friederici J. of cognitive.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
Neurophysiologic correlates of cross-language phonetic perception LING 7912 Professor Nina Kazanina.
The ERP Boot Camp Isolating Components with Difference Waves and
Neural Basis of the Ventriloquist Illusion Bonath, Noesselt, Martinez, Mishra, Schwiecker, Heinze, and Hillyard.
Introduction Can you read the following paragraph? Can we derive meaning from words even if they are distorted by intermixing words with numbers? Perea,
Multimodal Virtual Environments: Response Times, Attention, and Presence B 陳柏叡.
Video Games and Working Memory Derek M. Ellis Chris Blais Gene A. Brewer Department of Psychology Arizona State University The Entertainment Software Rating.
Introduction Ruth Adam & Uta Noppeney Max Planck Institute for Biological Cybernetics, Tübingen Scientific Aim Experimental.
Multimodal Perception
A direct comparison of Geodesic Sensor Net (128-channel) and conventional (30-channel) ERPs in tonal and phonetic oddball tasks Jürgen Kayser, Craig E.
RESEARCH QUESTIONS Might having to lie still without moving, or having to lie down rather than sit up, change the pattern of neural activity in very young.
Cortical Event-Realated Potentials to Auditory Stimuli 초고주파 및 항공전자통신 연구실 석사 2 차 : 임의선 (林宜宣) Lin Yixuan
Event-Related Potentials Chap2. Ten Simple Rules for Designing ERP Experiments (2/2) 임원진
Melanie Boysen & Gwendolyn Walton
[Ran Manor and Amir B.Geva] Yehu Sapir Outlines Review
S. Kramer1, K. Tucker1, A.L. Moro1, E. Service1, J.F. Connolly1
Attention Components and Creative Potential: An ERP Exploration
Neurofeedback of beta frequencies:
The Development of Emotional Interactions Across the Senses:
Viorica Marian, Sayuri Hayakawa, Tuan Q. Lam, & Scott Schroeder
Cognitive Brain Dynamics Lab
Michael S Beauchamp, Kathryn E Lee, Brenna D Argall, Alex Martin 
Performance pressure through social and monetary goals enhances speech category learning 10/10/17.
Multisensory Perception and Biological Motion
Multisensory integration: perceptual grouping by eye and ear
Volume 28, Issue 5, Pages e3 (March 2018)
Michael S Beauchamp, Kathryn E Lee, Brenna D Argall, Alex Martin 
Audiovisual Integration of Letters in the Human Brain
Novitskiy et al. Neuromath 2009
Michael J. Frank, Brion S. Woroch, Tim Curran  Neuron 
Christoph Kayser, Nikos K. Logothetis, Stefano Panzeri  Current Biology 
Shadowing Task Cherry, 1953 Attended Unattended
Presentation transcript:

Visual speech speeds up the neural processing of auditory speech van Wassenhove, V., Grant, K. W., & Poeppel, D. (2005) Proceedings of the National Academy of Sciences, 102(4), Jaimie Gilbert Psychology 593 October 6, 2005

Audio-Visual Integration Information from one modality (e.g., visual) can influence the perception of information presented in a different modality (e.g., auditory) Speech in noise McGurk Effect

Demonstration of McGurk Effect Audiovisual Speech Web-Lab -index.html Arnt Maasø University of Oslo McGurk_english.html

Unresolved questions about AV integration Behavioral evidence exists for vision altering the perception of speech, but… When does it occur in processing? How does it occur?

ERPs can help answer the “when” question EEG/MEG studies have demonstrated AV integration effects using oddball/mismatch paradigms These effects occur around ms A non-speech ERP study with non-ecologically valid stimuli demonstrated earlier interaction effects (40-95 ms) (Giard & Peronnet, 1999) Does AV integration for speech occur earlier than ms?

There’s a debate about the “how” question… Enhancement Audio-visual integration generates activity at multi-sensory integration sites, information possibly fed back to sensory cortices VS. Suppression Reduction of stimulus uncertainty by two corresponding sensory stimuli reduces the amount of processing required

The Experiments 3 experiments were conducted Each had behavioral and EEG measures Behavioral: Forced choice task EEG: Auditory P1/N1/P2 26 participants Experiment 1: 16 Experiment 2: 10 Experiment 3: 10 (of the 16 who participated in Experiment 1)

The Stimuli Audio /pa/ Audio /ta/ Audio /ka/ Visual /pa/ Visual /ta/ Visual /ka/ AV /pa/ AV /ta/ AV /ka/ Incongruent AV with Audio /pa/ + Visual /ka/ 1 Female face & voice for all stimuli In Exp. 1 & 2, each stimuli presented 100 times; total of 1000 trials

Experiment 1 Exp. 1 Stimuli presented in blocks of audio, or blocks of visual, or blocks of AV (congruent and incongruent) Participants knew before each block which stimuli were going to be presented

Experiment 2 Exp. 2 Stimuli presented in randomized blocks containing all stimuli types (A, V, Congruent AV, Incongruent AV) to reduce expectancy Task for both experiments: choose which stimuli was presented; for AV--choose what was heard while looking at the face

Experiment 3 Presented 200 Incongruent AV stimuli Task: choose what syllable you saw, neglect what you heard In all experiments, correct response to Incongruent AV = /ta/

Waveform Analysis Retained 75-80% of recordings after Artifact Rejection and Ocular Artifact Reduction Only correct responses were analyzed 6 electrodes used in analysis: FC3, FC4, FCz, CPz, P7, P8 Reference electrodes: Linked mastoids

Results This study’s answer to “How” Suppression/Deactivation Hypothesis AV N1 & P2 amplitude were significantly reduced compared to Auditory-alone peaks Performed separate analysis to determine if summing the responses to unimodal stimuli would result in the amplitude reduction present in the data—this was not the case; therefore the AV waveform is not a superposition of the 2 sensory waveforms, but reflects actual multisensory interaction.

Results: Experiment 1 N1/P2 Amplitude AV < A (p <.0001) N1/P2 Latency AV < A (significant, but confounded by interaction) Modality x Stimulus Identity P < T < K (p <.0001) Latency effect more pronounced in P2, but can occur as early as N1

Results: Experiment 2 N1/P2 Amplitude AV < A (p <.0001) N1/P2 Latency AV < A (p <.0001) Modality x Stimulus Identity (p <.06)

Results: comparison of Exp. 1 & Exp. 2 Similar results for Exp. 1 & 2; Temporal facilitation varied by Stimulus Identity but amplitude reduction did not; No evidence for attention effect (i.e., for expectancy affecting waveform morphology)

Temporal facilitation depends on visual saliency/signal redundancy More temporal facilitation is expected to occur if: The audio and the visual signals are redundant The visual cue (which naturally precedes the auditory cue) is more salient (Figure 3)

Results: Experiment 3/Incongruent AV Stimuli Incongruent AV stimuli in Exp. 1 & 2: no temporal facilitation Amplitude reduction present and equivalent to reduction seen for Congruent AV stimuli Experiment 3: Both temporal facilitation and amplitude reduction occurred

Visual speech effects on auditory speech Perceptual ambiguity/salience of visual speech affects processing time of auditory speech Incorporating visual speech with auditory speech reduces the amplitude of N1/P2 “independent of AV congruency, participant’s expectancy, and attended modality” (p. 1184)

Ecologically valid stimuli Suggest that AV speech processing is different from general multisensory integration due to the ecological validity of speech

Possible explanation for amplitude reduction Visemes provide information regarding place of articulation If this information is salient and/or redundant with auditory place of articulation cues (e.g., 2 nd and 3 rd formants), the auditory cortex does not need to analyze these frequency regions, resulting in fewer firing neurons

Analysis-by-Synthesis Model of AV Speech Perception Visual speech activates internal representation/prediction This representation/prediction is updated as more visual information is received over time Representation/prediction is compared to the incoming auditory signal Residual errors to this matching process are reflected by temporal facilitation and amplitude reduction effects Attended modality can influence temporal facilitation

Suggest 2 time scales for AV integration 1: feature stage 25 ms Latency facilitation (sub-)segmental analysis 2: perceptual unit stage 200 ms Amplitude reduction Syllable level analysis Independent of feature content and attended modality

Summary AV speech interaction occurs by the time N1 is elicited ( ms) Processing time of auditory speech varies by the saliency/ambiguity of visual speech Amplitude of AV ERP reduced when compared to amplitude of A-alone ERP

Questions Dynamic visual stimulus and ocular artifact If effects of AV integration are influenced by attended modality, would modality dominance also influence these effects? Are incongruent AV/McGurk stimuli ecologically valid?