The Neural Basis of Speech Perception – a view from functional imaging Sophie Scott Institute of Cognitive Neuroscience, University College London.

Slides:



Advertisements
Similar presentations
Figure Three-dimensional reconstruction of the left hemisphere of the human brain showing increased activity in ventrolateral area 45 during verbal.
Advertisements

Chapter 49 Language and Communication Copyright © 2014 Elsevier Inc. All rights reserved.
Chapter 4 Author: Brudzynski. 2 Fig. 1. Pathways supporting vocal communication in humans and monkeys. The pathways for parallel brain processing are.
Investigating motor-related sounds in the brain Zarinah Agnew, Carolyn McGettigan, Sophie Scott UCL Institute of Cognitive Neuroscience.
BUCNI Meeting BUCNI Meeting Aug Effects of spectral detail and tonal variation on speech intelligibility Kyong, Scott, Eisner and Rosen.
Acoustic versus perceptual accounts of speaker differentiation within anterior auditory cortex Nicolas J. Abreu Carolyn McGettigan Sophie K. Scott Institute.
Delayed auditory feedback: a study into vocal motor patterns UCL Institute of Cognitive Neuroscience Dr Zarinah Agnew Dr Carolyn McGettigan Briony Banks.
Auditory Forebrain Auditory Thalamus Auditory Cortex Auditory Projections to Polysensory Cortex Plasticity.
Chapter 12: Speech and Music Perception
Brain Structures Differ between Musicians and Non-Musicians
Emmanuel A Stamatakis Centre for Speech, Language and the Brain, Department of Experimental Psychology, University of Cambridge School of Psychological.
Central Auditory Pathways, Functions, and Language Central Auditory Pathways, Functions, and Language.
Chapter 12 Speech Perception. Animals use sound to communicate in many ways Bird calls Bird calls Whale calls Whale calls Baboons shrieks Baboons shrieks.
Human Communication.
Ch13. Biological Foundations of Language
The Neuroscience of Language. What is language? What is it for? Rapid efficient communication – (as such, other kinds of communication might be called.
Copyright © 2006 Pearson Education, Inc., publishing as Benjamin Cummings Central Nervous System (CNS)  CNS = Brain + spinal cord  Surface anatomy includes.
Speech Perception Overview of Questions Can computers perceive speech as well as humans? Does each word that we hear have a unique pattern associated.
Neural correlates of continuous and categorical sinewave speech perception: An FMRI study Rutvik Desai, Einat Liebenthal, Eric Waldron, and Jeffrey R.
Functional Anatomy of Spoken Input Note that the low-level auditory pathway is not specialized for speech sounds – Both speech and non-speech sounds activate.
Language. Using Language What is language for? Using Language What is language for? – Rapid, efficient communication To accomplish this goal, what needs.
Read this article for Friday next week [1]Chelazzi L, Miller EK, Duncan J, Desimone R. A neural basis for visual search in inferior temporal cortex. Nature.
Structure and function
Results Animals with recognition displayed increased BDNF in the iTC, CA1 of the hippocampus, the diagonal band, basolateral amygdala and the anterior.
Auditory-acoustic relations and effects on language inventory Carrie Niziolek [carrien] may 2004.
Auditory Neuroscience - Lecture 6 Hearing Speech auditoryneuroscience.com/lectures.
Lateralization & The Split Brain and Cortical Localization of Language.
Auditory Processing Physical DimensionPerceptual Dimension AmplitudeLoudness FrequencyPitch ComplexityTimbre.
Recent Findings in the Neurobiology & Neuropsychology of Reading Processes A. Maerlender, Ph.D. Clinical School Services & Learning Disorders.
By Peter Dang and Kyle Dennison.  Ability to acquire language is a unique and essential human trait  Chomsky’s Universal Grammar The human brain contains.
Central Nervous System (CNS) CNS –brain –spinal cord.
Chapter 11: Perception of Sound Overview of Questions If a tree falls in the forest and no one is there to hear it, is there a sound? What is it that makes.
Comparative Diffusion Tensor Imaging (DTI) Study of Tool Use Pathways in Humans, Apes and Monkeys Ashwin G. Ramayya 1,2, Matthew F. Glasser 1, David A.
Hearing in Time Slides used for talk to accompany Roger go to yellow three … Sarah Hawkins* and Antje Heinrich** *Centre for Music and Science, University.
Sensorimotor systems Chapters 8.
Background The physiology of the cerebral cortex is organized in hierarchical manner. The prefrontal cortex (PFC) constitutes the highest level of the.
Chapter 12: Auditory Localization and Organization
Short-Term Reorganization of Auditory Analysis Induced by Phonetic Experience Liebenthal et al. (2003). JoCN. Audrey Kittredge 593: Neuroimaging of Language.
Coricelli and Nagel (2008) Introduction Methods Results Conclusion.
Basic Pattern of the Central Nervous System Spinal Cord – ______________________________ surrounded by a _ – Gray matter is surrounded by _ myelinated.
Four lobes of the cerebral cortex FRONTAL LOBE OCCIPITAL LOBE TEMPORAL LOBE PARIETAL LOBE.
Chapter 13: Speech Perception. The Acoustic Signal Produced by air that is pushed up from the lungs through the vocal cords and into the vocal tract Vowels.
Auditory Neuroscience 1 Spatial Hearing Systems Biology Doctoral Training Program Physiology course Prof. Jan Schnupp HowYourBrainWorks.net.
When the Brain is attending a cocktail party When the Brain is attending a cocktail party Rossitza Draganova.
Human Anatomy & Physiology FIFTH EDITION Elaine N. Marieb PowerPoint ® Lecture Slide Presentation by Vince Austin Copyright © 2003 Pearson Education, Inc.
Acoustic Illusions & Sound Segregation Reading Assignments for Quizette 3 Chapters 10 & 11.
Perception of a sensory experience can change even though the input remains the same.
Group 4 Alicia Iafonaro Anthony Correa Baoyu Wang Isaac Del Rio
The Relation Between Speech Intelligibility and The Complex Modulation Spectrum Steven Greenberg International Computer Science Institute 1947 Center Street,
Articulatory Net I.2 Oct 14, 2015 – DAY 21
Orienting Attention to Semantic Categories T Cristescu, JT Devlin, AC Nobre Dept. Experimental Psychology and FMRIB Centre, University of Oxford, Oxford,
Fundamentals of Sensation and Perception
Introduction  Recent neuroimaging studies of memory retrieval have reported the activation of a medial and left – lateralised memory network that includes.
Sundeep Teki Wellcome Trust Centre for Neuroimaging University College London, UK Auditory figure-ground segregation in complex acoustic scenes.
By: Angela D. Friederici Presented By: Karol Krzywon.
Cross modal and linguistic interactions in speech perception
Dept. of Brain & Cognitive Sciences
Auditory Localization in Rooms: Acoustic Analysis and Behavior
Risto Näätänen University of Tartu, Estonia
Is There a Preference for Low-Level Linguistic Processing in Autism?
Auditory Neuroscience - Lecture 6 Hearing Speech
Volume 56, Issue 6, Pages (December 2007)
Neural Changes following Remediation in Adult Developmental Dyslexia
Volume 30, Issue 3, Pages (May 2001)
Experimental Design in Functional Neuroimaging
UNIT 3 THE CONSCIOUS SELF
Neural Changes following Remediation in Adult Developmental Dyslexia
Liping Wang, Lynn Uhrig, Bechir Jarraya, Stanislas Dehaene 
The Prefrontal Cortex—An Update
César F. Lima, Saloni Krishnan, Sophie K. Scott 
Presentation transcript:

The Neural Basis of Speech Perception – a view from functional imaging Sophie Scott Institute of Cognitive Neuroscience, University College London

This approach to speech perception Speech is an auditory signal It is possible to address the neural processing of speech within the framework of auditory cortical processing. This is not synonymous with the entire language system. If one is a skilled speaker of a language, then speech perception is obligatory.

Functional imaging Where neural activity occurs, blood is directed. Measure neural activity by tracking these changes in local blood flow. Thus measuring mass synaptic activity Poor temporal resolution Essentially a comparison of blood flow changes across conditions - so the baseline comparisons are critical

Listening Wise et al, Lancet, 2001

Neuroanatomy of speech Speech production Speech perception

sts Ins TS2 TS1 TS3 A1 Tpt paAlt Pro caudal CORE BELT PARABELT rostral medial lateral sts dorsal RTM RM CM R RT A1 RTL AL ML MC RP CP Scott and Johnsrude, 2003, from Romanski et al, 1999

AIRRTCLMLALSTGc CBP RBP STGr Dorsal prearcuate (8a) Dorsal principal sulcus (46) Inferior convexity (12) Orbital polar From Kaas and Hackett, 1999 CoreBeltParabelt Prefrontal cortex

tonotopy bandwidth Conspecific vocalisations Spatial representations

HG PT C B PB Assoc Tpt AnteriorPosterior Ventral STS STP STS STP Human Monkey

medial Scott and Johnsrude, 2003 anterior posterior lateral AA MA LA A1 PA STA ALA LP

medial Scott and Johnsrude, 2003 anterior posterior lateral Sounds with harmonic structure against pure tones: Hall, Johnsrude et al., 2002 Frequency modulated tones against unmodulated tones: Hall, Johnsrude et al., 2002 Amplitude modulated noise against unmodulated noise: Giraud et al, 1999 Spectral change against steady state sounds: Thivard et al, 2000

Hierarchical processing Structure in sound is computed beyond primary auditory cortex More complex structure (e.g. spectral change) processed further from PAC How does this relate to speech processing?

speech noise vocoded speech rotated speech rotated noise vocoded speech

Sp VCo RSp RVCo Anterior Z = Z = Z = Z = 5.7 (Sp + VCo + RSp) - RVCo (Sp + VCo) - (RSp + RVCo) Left hemisphere Scott, Blank, Rosen and Wise, 2000

1 0 2 Sp VCo RSp RVCo Right hemisphere Anterior Z = 6.7 (Sp + RSp) - (VCo + RVCo) Scott, Blank, Rosen and Wise, 2000

Intelligibility

Plasticity within this system Naïve subjects were scanned before they could understand noise vocoded speech, then they were trained, then scanned again.

Activity to noise vocoded speech after a training period, relative to prior activity to NVC before the training period. Narain, Wise, Rosen, Matthews, Scott, under review. Flexibility in speech perception: learning to understand noise vocoded speech As well as left lateralised STS, there is involvement of left premotor cortex and the left anterior thalamus (which receive projections from the belt and parabelt).

Spectrograms of the stimuli (speech) (rotated speech) 16R 3R

Intelligibility - behavioural data

Right Scott, Rosen, Lang and Wise, 2006 Z=5.96 x=64 y=-4 z=-2 Z=4.73 x=-48 y=-16 z=-16 Z=4.52 x=-64 y=-28 z=8Z=5.6 x=-62 y=-10 z=80 Left R 16R

medial Scott and Johnsrude, 2003 anterior posterior lateral Sounds with harmonic structure against pure tones: Hall, Johnsrude et al., 2002 Frequency modulated tones against unmodulated tones: Hall, Johnsrude et al., 2002 Amplitude modulated noise against unmodulated noise: Giraud et al, 1999 Spectral change against steady state sounds: Thivard et al, 2000 Peak responses to Intelligibility (Scott et al, 2006)

Speech specific processing Does not occur in primary auditory cortexd Begins early in auditory cortex - in areas that also respond to AM As we move forward down the STS, the responses become less sensitive to acoustic structure - resembles behavioural profile

Speech comprehension - The role of context e.g., words recognised more easily in sentences “The ship sailed the sea” > “Paul discussed the dive”. Can we identify the neural basis of this contextual modulation of speech comprehension? (Miller et al., 1951; Boothroyd and Nittrouer, 1988; Grant and Seitz, 2000; Stickney and Assmann, 2001; Davis et al., 2005)

(noise vocoding: Shannon et al., 1995 predictability: Kalikow et al., 1977)

Low predictability: log increase with more channels …‘Sue was interested in the bruise’… jonas obleser 27

Behav 2 low+high High predictability: influence at intermediate number of channels …‘He caught the fish in his net’… …‘Sue was interested in the bruise’… jonas obleser 28

(cf. e.g. Binder et al. 2000; Scott et al., 2000; Davis & Johnsrude 2003; Zekveld et al., 2006) Bottom-up processes: correlations with number of channels RFX p 30 Obleser, Wise, Dresner, & Scott, 2007

Left-hemispheric array of brain regions when context affects comprehension Lateral Prefrontal (BA 8) Posterior Cingulate (BA 30) Medial Prefrontal (BA 9) Angular Gyrus (BA 39) Ventral IFG (BA 47) RFX p 30 Obleser, Wise, Dresner, & Scott, 2007

findings A range of brain areas outwith auditory cortex contribute to ‘top down’ semantic influences on speech perception Further studies will be able to dissociate the contributions of different linguistic factors

Words are not the only things we say

Non speech sounds? x=54 Regions in red respond to noises and rotated noises Regions in yellow respond to noises and rotated noises

1 0 2 Sp VCo RSp RVCo Right hemisphere Anterior Z = 6.7 (Sp + RSp) - (VCo + RVCo)

What drives lateral asymmetry? Previous studies have not generally used ‘speech-like’ acoustic modulations We aimed to manipulate speech stimuli to vary the amplitude and spectral properties of speech independently Control for intelligibility Do we see additive effects of amplitude and spectral modulations? Are these left lateralised?

Steady spectrum, steady amplitude Steady spectrum, varying amplitude Varying spectrum, steady amplitude Varying spectrum, varying amplitude

Ideal additive effects Effect size Down for flat amplitude and spectrumSimilar response to AM and SpM Significantly more activated by stimuli with both AM and SpM

Additive effects Flat AM SpM SpMAM PET scanning, 16 runs, N=13, thresholded at p<0.0001, 40 voxels

Additive effects Flat AM SpM SpMAM PET scanning, 16 runs, N=13, thresholded at p<0.0001, 40 voxels

But… Is there a problem - were these stimuli really processed as speech? To address this, 6 of the 13 subjects were pretrained on speech exemplars, and the speech stimuli were included as a 5th condition.

A B C D E speech

A B C D E

Speech conditions Flat AM SpM SpMAM N=6, thresholded at p<0.0001, 40 voxels

Speech conditions Flat AM SpM SpMAM N=6, thresholded at p<0.0001, 40 voxels

Asymmetries in speech perception Exist! Are not driven by simple properties of the speech signal Right - preferentially processes speech- like sounds - voices? Left - processes linguistically relevant information

Posterior auditory areas In primates, medial posterior areas show auditory and tactile responses What do these areas do in speech processing in humans?

Speaking and mouthing This region, in the left posterior temporal- parietal junction, responds when subject repeat a phrase, mouth the phrase silently, or go ‘uh uh’, over mentally rehearsing the phrase Wise, Scott, Blank, Murphy, Mummery and Warburton, 2001 Wise et al, 2001, Brain

Listening over silence Amount of DAF (0, 50, 125, 200ms)

DAF peak on right

Neural basis of speech perception Hierarchical processing of sound in auditory cortex The anterior ‘what’ pathway is important in the perceptual processing of speech Activity in this system can be modulated by top down linguistic factors There are hemispheric asymmetries in speech perception - the left is driven by phonetic, lexical and linguistic properties: the right is driven by pitch variation, emotion and indexical properties There are sensory motor links in posterior auditory areas - part of a ‘how’ pathway?

what where what how where Scott, in press Scott, Current Opinions in Neurobiology, 2005

Charlotte Jacquemot Frank Eisner Disa Sauter Carolyn McGettigan Narly Golestani Jonas Obleser Sophie Scott Stuart Rosen Richard Wise Charvy Narain Andrew Faulkner Hideki Takaso