Cross-modal Prediction in Speech Perception Carolina Sánchez, Agnès Alsius, James T. Enns & Salvador Soto-Faraco Multisensory Research Group Universitat.

Slides:



Advertisements
Similar presentations
Listening skills: how to untangle the noise and find the message Judy Copage.
Advertisements

Electrophysiology of Visual Attention. Does Visual Attention Modulate Visual Evoked Potentials? The theory is that Visual Attention modulates visual information.
English only VS. L1 support Looking at English language acquisition of native Spanish speaking children.
Visual speech speeds up the neural processing of auditory speech van Wassenhove, V., Grant, K. W., & Poeppel, D. (2005) Proceedings of the National Academy.
Human Speech Recognition Julia Hirschberg CS4706 (thanks to John-Paul Hosum for some slides)
Modality-specific interaction between phonology and semantics Gail Moroschan & Chris Westbury Department of Psychology, University of Alberta, Edmonton,
Language and Cognition Colombo, June 2011 Day 8 Aphasia: disorders of comprehension.
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Spoken Word Recognition 1 Language Use and Understanding.
Ling 240: Language and Mind Acquisition of Phonology.
9/22/10Psyc / Ling / Comm 525 Fall 10 Semantic Priming (Phenomenon & Tool)...armkitchentree Related prime >doctoractor < Unrelated prime nurse floor...
Experimental Psychology PSY 433 Chapter 7 Perception (Cont.)
The Perception of Speech. Speech is for rapid communication Speech is composed of units of sound called phonemes –examples of phonemes: /ba/ in bat, /pa/
Auditory Word Recognition
Background Dissociation: ◦ Lexical-gender (king) - recovered directly from the lexicon ◦ Stereotypical-gender (minister) – inferred from pragmatic information.
Working Memory: The Feature Model Presented by: Umer Fareed.
Does sentence constraint influence word recognition in bilinguals? Evidence from Event-Related Potentials and RTs Pascal E. A. Brenders, Janet G. van Hell,
Exam 1 Monday, Tuesday, Wednesday next week WebCT testing centre Covers everything up to and including hearing (i.e. this lecture)
Read this article for Friday [1]Chelazzi L, Miller EK, Duncan J, Desimone R. A neural basis for visual search in inferior temporal cortex. Nature 1993;
Syntactic Priming in Bilinguals: Effects of verb repetition in an L2-monolingual and cross-lingual setting Sofie Schoonbaert 1, Robert Hartsuiker 1, &
Memory III Working Memory & Brain. Atkinson & Shiffrin (1968) Model of Memory.
Interactions between Stimulus- Specific Adaptation and Visual Auditory Integration in the Forebrain of the Barn Owl Amit Reches, Shai Netser, and Yoram.
Word Recognition By: S. Jay Samuels Michele Farah & Marsha Malone.
PSY 369: Psycholinguistics
Memory III Working Memory. Atkinson & Shiffrin (1968) Model of Memory.
The Perception of Speech
The auditory and the visual percept evoked by the same audiovisual stimuli Hartmut Traunmüller Niklas Öhrström Dept. of Linguistics, University of Stockholm.
The role of auditory-visual integration in object recognition Clara Suied 1, Nicolas Bonneel 2 and Isabelle Viaud-Delmon 1 1 CNRS – UPMC UMR 7593 Hôpital.
Experimental study of morphological priming: evidence from Russian verbal inflection Tatiana Svistunova Elizaveta Gazeeva Tatiana Chernigovskaya St. Petersburg.
Graziella Quattrocchi & Louise Marshall Methods for Dummies 2014
1 Sensory and Short-Term Memory PSY 421 – Fall 2004.
Results 1.Boundary shift Japanese vs. English perceptions Korean vs. English perceptions 1.Category boundary was shifted toward boundaries in listeners’
Acoustic Cues to Laryngeal Contrasts in Hindi Susan Jackson and Stephen Winters University of Calgary Acoustics Week in Canada October 14,
Growing up Bilingual: One System or Two? Language differentiation and speech perception in infancy.
Right hemisphere sensitivity to word & sentence level context: Evidence From Event-Related Brain Potentials. Coulson, S. Federmeier, K.D., Van Petten,
SPEECH PERCEPTION DAY 16 – OCT 2, 2013 Brain & Language LING NSCI Harry Howard Tulane University.
Understanding Action Verbs- Embodied Verbal Semantics Approach Pavan Kumar Srungaram M.Phil Cognitive Science (09CCHL02) Supervisor: Prof. Bapi.
Pragmatically-guided perceptual learning Tanya Kraljic, Arty Samuel, Susan Brennan Adaptation Project mini-Conference, May 7, 2007.
TEMPORAL DISPLACEMENTS: NEW EVIDENCE ON THE BINDING PROBLEM Ekaterina Pechenkova Moscow State University.
1 Cross-language evidence for three factors in speech perception Sandra Anacleto uOttawa.
The Source of Enhanced Cognitive Control in Bilinguals: Evidence From Bimodal-Bilinguals Gigi Luk 1, Jennie Pyers 2, Karen Emmorey 3 & Ellen Bialystok.
Human Abilities 2 How do people think? 1. Agenda Memory Cognitive Processes – Implications Recap 2.
Introduction Can you read the following paragraph? Can we derive meaning from words even if they are distorted by intermixing words with numbers? Perea,
Phonological Priming and Lexical Access in Spoken Word Recognition Christine P. Malone Minnesota State University Moorhead.
Alejandro Peréz, Margaret Gillon Dowens, Nicola Molinaro, Yasser Iturria-Medina, Paulo Barraza, Lorna García-Pentón, and Manuel Carreiras.
Articulatory Net I.2 Oct 14, 2015 – DAY 21
The Source of Enhanced Cognitive Control in Bilinguals: Evidence From Bimodal-Bilinguals Gigi Luk 1, Jennie Pyers 2, Karen Emmorey 3 & Ellen Bialystok.
Tonal Violations Interact with Lexical Processing: Evidence from Cross-modal Priming Meagan E. Curtis 1 and Jamshed J. Bharucha 2 1 Dept. of Psych. & Brain.
Working memory and second language learning Graham Hitch, Meesha Warmington & Swathi Kandru.
“Optimal learning occurs when the two hemispheres of the brain work together. Any teaching strategy such as music that integrates the function of both.
Methods Identifying the Costs of Auditory Dominance on Visual Processing: An Eye Tracking Study Wesley R. Barnhart, Samuel Rivera, & Christopher W. Robinson.
Lab 4: Pseudo-homophones They sound like words, but they aren’t.
Richard A. Ambert Torres Dr. Charles Nagle – Faculty Mentor
Feature Binding: Not Quite So Pre-attentive Erin Buchanan and M
Assist. Prof. Dr. Ilmiye Seçer Fall
Phonological Priming and Lexical Access in Spoken Word Recognition
From: Cross-modal attention influences auditory contrast sensitivity: Decreasing visual load improves auditory thresholds for amplitude- and frequency-modulated.
The toolbox for language description Kuiper and Allan 1.2
Multisensory perception: Beyond modularity and convergence
XinTONG yu ALEX BATES SEPTEMBER 12TH, 2017
SECOND LANGUAGE LISTENING Comprehension: Process and Pedagogy
بسمه تعالی کارگاه ارزشیابی پیشرفت تحصیلی
Phonological Priming and Lexical Access in Spoken Word Recognition
Cognitive profile of higher education students with dyslexia
Multisensory Perception and Biological Motion
Multisensory integration: perceptual grouping by eye and ear
Human vision: function
Multisensory Integration: Maintaining the Perception of Synchrony
Shadowing Task Cherry, 1953 Attended Unattended
Presentation transcript:

Cross-modal Prediction in Speech Perception Carolina Sánchez, Agnès Alsius, James T. Enns & Salvador Soto-Faraco Multisensory Research Group Universitat Pompeu Fabra Barcelona

Auditory + visual performance MSI enhancement Background Visual + Auditory Improve Speech Perception Multisensory Integration

Background Prediction within one sensory modality Many levels of information processing –Phonological prediction “ This morning I went to the library and borrowed a … book” (De Long, 2005; Pickering, 20707) –Visual prediction: Visual search (Enns, 2008; Dambacher, 2009) –Sensorimotor prediction: forward model (Wolpert, 1997)

Predictive coding Pickering, 2007

Hypothesis If there exists prediction within the same modality, and if predictive coding models can account for prediction at a phonological level, then … Predictive Coding could occur across different sensory modalities too.

Indirect evidences of cross-modal transfer in speech van Wassenhove’s, 2005 time ERPs Amplitud reduction Shortening latency /pa/ high visual saliency /ka/ short visual saliency

Our study Visual prediction Auditory prediction Visual-to-auditory cross-modal prediction Auditory-to-visual cross-modal prediction

Visual prediction Visual stream Auditory stream V A With visual informative visual context Without informative context Task : AV Match vs. AV Mismatch Target fragment Context fragment speech non speech

Results * Reaction time msec matchmismatch With visual informative context Without informative context * With previous context participants respond faster than without it. VISUAL PREDICTION

Auditory prediction Visual stream Auditory stream V A With auditory informative auditory context Without informative context speech non speech Task : AV Match vs. AV Mismatch Target fragment Context fragment

Results * With auditory informative context Without informative context Reaction time msec match mismatch * With previous context participants respond faster than without it. AUDITORY PREDICTION

Visual vs. Auditory Visual prediction Auditory prediction Rts msec congruentincongruent With visual informative context Without informative context * With auditory informative context Without informative context Rts msec congruent incongruent *

Conclusions Visual prediction Auditory prediction Is this prediction cross-modal?

Predictability of Vision-to-Audition Design of the experiment V A Mismatch Unimodal continued Auditory stream Visual stream Match Unimodal continued V A Discontinued Match V A Discontinued Mismatch V A Cross-modal continued Mismatch

Predictability of Vision-to-Audition Stimuli V A Mismatch V A V A Unimodal continued Discontinued Cross-modal continued

Results Participants were faster in the cross-modal condition than in the completely incongruent one. VISUAL –TO-AUDITORY PREDICTION Reaction time msec * Visual Auditory Unimodal continued Discontinued Cross-modal continued

Predictability of Audition-to-Vision Design of the experiment Auditory stream Visual stream Match Unimodal continued V A Mismatch Unimodal continued V A Match Discontinued V A Mismatch Discontinued V A Mismatch Cross-modal continued

Reaction time msec Visual Auditory Unimodal continued Discontinued Cross-modal continued Results We didn’t find any difference between the mismatch condicions NO AUDITORY-TO-VISUAL PREDICTION

Conclusions There is some kind of prediction from vision-to- auditory modality There is not any prediction from auditory-to-vision modality Does this prediction depend on the language?

Canadian participants with english sentences VISUAL –TO-AUDITORY PREDICTION IN NATIVE LANGUAGE Reaction time msec * Visual Auditory Unimodal continued Discontinued Cross-modal continued Reaction time msec * Visual Auditory Unimodal continued Discontinued Cross-modal continued Spanish participants with spanish sentences Results (L1)

Canadian participants with english sentences Reaction time msec No differences between the mismatch conditions No prediction from auditory-to- visual modality in native language Spanish participants with spanish sentences Reaction time msec Visual Auditory Unimodal continued Discontinued Cross-modal continued Visual Auditory Unimodal continued Discontinued Cross-modal continued

Conclusions There is some kind of prediction from vision-to- auditory modality in L1 There is not any prediction from auditory-to-vision modality L1 What happens with an unknown language?

Unknown language : visual to auditory Canadian participants with spanish sentences NO VISUAL-TO-AUDITORY IN OTHER LANGUAGE Reaction time msec Visual Auditory Unimodal continued Discontinued Cross-modal continued

Unknown language: auditory to visual Spanish participants with english sentences Canadian participants with spanish sentences Reaction time msec Reaction time msec No differences between the mismatch conditions No prediction from auditory-to- visual modality in other language Visual Auditory Unimodal continued Discontinued Cross-modal continued Visual Auditory Unimodal continued Discontinued Cross-modal continued

Conclusions No visual-to-auditory cross-modal prediction in an unknown language… it seems that some level of knowledge about the articulatory phonetics of the language is required to obtain the advantage of the predictive coding No auditory-to-visual cross-modal prediction

General Conclusions Unimodal prediction from visual to visual modality from auditory to auditory L1: ASYMMETRY –Cross-modal prediction from visual-to-auditory modality –No cross-modal prediction from auditory-to-visual modality Unknown language: previous knowledge of the language is neccesary to make the prediction –No cross-modal prediction from visual-to-auditory modality –No cross-modal prediction from auditory-to-visual modality

-Agnès Alsius, Postdoc Queen’s University -Antonia Najas, MA/ Research Assistant Universitat Pompeu Fabra -Phil Jaekl, Postdoc Universitat Pompeu Fabra - All the people of the Vision Lab, UBC, Vancouver Thanks to… Thanks for your attention!!