Alan Kan, Corey Stoelb, Matthew Goupell, Ruth Litovsky

Slides:



Advertisements
Similar presentations
Auditory Localisation
Advertisements

Perception Chapter 11: Hearing and Listening
Binaural Hearing Or now hear this! Upcoming Talk: Isabelle Peretz Musical & Non-musical Brains Nov. 12 noon + Lunch Rm 2068B South Building.
3-D Sound and Spatial Audio MUS_TECH 348. Psychology of Spatial Hearing There are acoustic events that take place in the environment. These can give rise.
Properties of Sound EQ: How does intensity, loudness, frequency and pitch affect sound waves?
Localizing Sounds. When we perceive a sound, we often simultaneously perceive the location of that sound. Even new born infants orient their eyes toward.
Vocal Emotion Recognition with Cochlear Implants Xin Luo, Qian-Jie Fu, John J. Galvin III Presentation By Archie Archibong.
Source Localization in Complex Listening Situations: Selection of Binaural Cues Based on Interaural Coherence Christof Faller Mobile Terminals Division,
King Chung, Ph.D., CCC-A, FAAA Amplification and Communication Research Lab ACRL.
Localization cues with bilateral cochlear implants Bernhard U. Seeber and Hugo Fastl (2007) Maria Andrey Berezina HST.723 April 8 th, 2009.
Using Fo and vocal-tract length to attend to one of two talkers. Chris Darwin University of Sussex With thanks to : Rob Hukin John Culling John Bird MRC.
ICA Madrid 9/7/ Simulating distance cues in virtual reverberant environments Norbert Kopčo 1, Scott Santarelli, Virginia Best, and Barbara Shinn-Cunningham.
Interrupted speech perception Su-Hyun Jin, Ph.D. University of Texas & Peggy B. Nelson, Ph.D. University of Minnesota.
Effect of roving on spatial release from masking for amplitude-modulated noise stimuli Norbert Kopčo *, Jaclyn J. Jacobson, and Barbara Shinn-Cunningham.
Auditory Objects of Attention Chris Darwin University of Sussex With thanks to : Rob Hukin (RA) Nick Hill (DPhil) Gustav Kuhn (3° year proj) MRC.
Hearing & Deafness (3) Auditory Localisation
Spectral centroid 6 harmonics: f0 = 100Hz E.g. 1: Amplitudes: 6; 5.75; 4; 3.2; 2; 1 [(100*6)+(200*5.75)+(300*4)+(400*3.2)+(500*2 )+(600*1)] / = 265.6Hz.
Sound source segregation (determination)
Acoustical Society of America, Chicago 7 June 2001 Effect of Reverberation on Spatial Unmasking for Nearby Speech Sources Barbara Shinn-Cunningham, Lisa.
1 Recent development in hearing aid technology Lena L N Wong Division of Speech & Hearing Sciences University of Hong Kong.
Different evaluations for different kinds of hearing Matthew B. Winn Au.D., Ph.D. Waisman Center, UW-Madison Dept. of Surgery.
Humans can hear sounds at frequencies from about 20Hz to 20,000Hz.
Auditory cortical monitoring prevents speech errors before they happen Caroline A. Niziolek UCSF Depts. of Radiology and Otolaryngology – Head and Neck.
Ruth Litovsky University of Wisconsin Madison, WI USA Brain Plasticity and Development in Children and Adults with Cochlear Implants
METHODOLOGY INTRODUCTION ACKNOWLEDGEMENTS LITERATURE Low frequency information via a hearing aid has been shown to increase speech intelligibility in noise.
Auditory transduction Figure by MIT OCW. After figure in: Bear, Mark F., Barry W. Connors, and Michael A. Paradiso. Neuroscience: Exploring the Brain.
1 Loudness and Pitch Be sure to complete the loudness and pitch interactive tutorial at … chophysics/pitch/loudnesspitch.html.
Sh s Children with CIs produce ‘s’ with a lower spectral peak than their peers with NH, but both groups of children produce ‘sh’ similarly [1]. This effect.
Sounds in a reverberant room can interfere with the direct sound source. The normal hearing (NH) auditory system has a mechanism by which the echoes, or.
Localization of Auditory Stimulus in the Presence of an Auditory Cue By Albert Ler.
creating sound value TM Spatial release from masking deficits in hearing-impaired people: Is inadequate audibility the problem? Helen.
‘Missing Data’ speech recognition in reverberant conditions using binaural interaction Sue Harding, Jon Barker and Guy J. Brown Speech and Hearing Research.
Sound Energy II Or Tell me more about those scales…
Hearing Research Center
Human Detection and Localization of Sounds in Complex Environments W.M. Hartmann Physics - Astronomy Michigan State University QRTV, UN/ECE/WP-29 Washington,
A Child with a Hearing Impairment, Including Deafness ECEA Disability Category, Definition and Eligibility Criteria March 2013.
Laboratory for Experimental ORL K.U.Leuven, Belgium Dept. of Electrotechn. Eng. ESAT/SISTA K.U.Leuven, Belgium Combining noise reduction and binaural cue.
Auditory Neuroscience 1 Spatial Hearing Systems Biology Doctoral Training Program Physiology course Prof. Jan Schnupp HowYourBrainWorks.net.
Hearing Sound and the limits to hearing Structure of the ear: Outer, middle, inner Outer ear and middle ear functions Inner ear: the cochlea - Frequency.
Humans can hear sounds at frequencies from about 20Hz to 20,000Hz.
Modeling human localization in multiple sound-source scenarios Jonas Braasch Jens Blauert Institut für Kommunikationsakustik Ruhr-Universität Bochum
Development of Sound Localization 2 How do the neural mechanisms subserving sound localization develop?
PSYC Auditory Science Spatial Hearing Chris Plack.
3-D Sound and Spatial Audio MUS_TECH 348. Environmental Acoustics, Perception and Audio Processing: Envelopment.
Fletcher’s band-widening experiment (1940)
The role of reverberation in release from masking due to spatial separation of sources for speech identification Gerald Kidd, Jr. et al. Acta Acustica.
What can we expect of cochlear implants for listening to speech in noisy environments? Andrew Faulkner: UCL Speech Hearing and Phonetic Sciences.
SPATIAL HEARING Ability to locate the direction of a sound. Ability to locate the direction of a sound. Localization: In free field Localization: In free.
3-D Sound and Spatial Audio MUS_TECH 348. What do these terms mean? Both terms are very general. “3-D sound” usually implies the perception of point sources.
Sound Localization and Binaural Hearing
What in the World is that Audiogram Saying?
Auditory Localization in Rooms: Acoustic Analysis and Behavior
PSYCHOACOUSTICS A branch of psychophysics
Precedence-based speech segregation in a virtual auditory environment
Petr Marsalek Prague, CZ, Charles University,
Copyright © American Speech-Language-Hearing Association
Speech recognition with amplitude and frequency modulations: Implications for cochlear implant design Fan-Gang Zeng Kaibao Nie Ginger Stickney Ying-Yee.
Consistent and inconsistent interaural cues don't differ for tone detection but do differ for speech recognition Frederick Gallun Kasey Jakien Rachel Ellinger.
Presenter: Prof.Dr.-Eng. Gheorghe-Daniel Andreescu
Copyright © American Speech-Language-Hearing Association
Copyright © American Speech-Language-Hearing Association
Correlation between Binaural Perception and Brainstem Lesions
RESULTS: Individual Data
Localizing Sounds.
3 primary cues for auditory localization: Interaural time difference (ITD) Interaural intensity difference Directional transfer function.
Properties of Sound EQ: How does intensity, loudness, frequency and pitch affect sound waves?
3-D Sound and Spatial Audio
Properties of Sound EQ: How does intensity, loudness, frequency and pitch affect sound waves?
Localization behavior in ITD-alone condition and head turn analysis.
Week 13: Neurobiology of Hearing Part 2
Presentation transcript:

Alan Kan, Corey Stoelb, Matthew Goupell, Ruth Litovsky Possible Implications of Interaural Mismatch in the Place-of-Stimulation  on Spatial Release from Masking in Cochlear Implant Listeners Alan Kan, Corey Stoelb, Matthew Goupell, Ruth Litovsky

Cochlear Implants (CI) restore hearing Cochlear implants are used to restore hearing in profoundly deaf individuals by converting acoustical sound into electrical signals that stimulate the auditory nerves and take advantage of the tonotopic organization of the cochlea Bilateral cochlear implantation has become much more common in the last 8-10 years. Source: http://www.cochlear.com/uk/node/6858

Binaural cues tells us where sounds come from Interaural time difference (ITD) Bang! Theoretically, with bilateral implantation, binaural cues such as ITDs (animate), which is the difference in the time of arrival of a sound at the 2 ears, and ILDs (animate), the difference in intensity at the two ears, become available to the cochlear implant user. These cues should ideally allow the bilateral user to tell where sounds are coming from on the horizontal plane. Research has shown that with 2 implants, bilateral users perform better at sound localization tasks and understanding speech in noisy environments over unilateral users, which you would expect since they have access to binaural cues. Interaural level difference (ILD)

Spatial Release from Masking (SRM) The improvement in speech-in-noise understanding gained from a separation between target and masker sources Masker Target One other benefit of bilateral implantation is that there should be a benefit of spatial release from masking or SRM since implantees can somewhat tell where sounds are coming from. SRM is the improvement in speech-in-noise understanding gained from a separation between target and masker sources.

Normal hearing listeners SRM (dB) CI users Normal hearing listeners This is a figure taken from Loizou et al. (2009) and it compares binaural listening in CI users and normal hearing listeners. I particularly want to focus on the data presented in the first row which shows the data for an SRM condition. Here, we see that SRM for CI users is on average about 2-4 dB, where as in NH listeners it is much higher at 6-8 dB. What we see here is that CI users seem to get a small amount of SRM but it is not as much as NH listeners. However, much of this SRM in CI users is due to head shadow rather than a true benefit of spatial release. There can be a few reasons why CI users show very little benefit of SRM. They include: Neural survival after deafness How well the acoustic sounds are being encoded and transmitted via electrical signals Surgical issues leading to differences in the insertion depths of the implants in the two cochlea. This last issue is the one I want to focus on for the rest of my talk. Loizou et. al. (2009)

Interaural place-of-stimulation mismatch Base Apex Electrode # 2 4 6 8 10 12 14 16 18 20 22 7000 Hz 4000 Hz Left 2 4 6 8 10 12 14 16 18 20 22 Electrode # Right Let me explain what I mean by an interaural mismatch and its implications Imagine two electrode arrays, one inserted into each cochlea but at different depths. Typically a CI speech processor divides acoustic information up into different frequency bands and the information in each band is sent to the electrode number allocated to that band. However, if the electrode arrays are not inserted at the same depth, it may excite a different region of the cochlea in each ear leading to an interaural mismatch in the acoustic information being sent to the brain for processing. Now imagine that the acoustic information contained cues for a sound source location, that is ITD and ILD cues, how would the brain interpret this information, particularly if frequency matching is important for ITD/ILD processing physiologically? The answer to this question might affect how well CI users can get an advantage from SRM. 20000 Hz 20 Hz

Mismatch affects SRM Here’s some SRM data that we’ve been collecting in NH listeners using a CI simulator and with different amounts of interaural mismatch. Here, I’m plotting SRM as a function of interaural mismatch shown in mm along the cochlea. What we see is that there is quite a bit of individual variation in terms of how much mismatch affects SRM. For SYV, we see that SRM performance is pretty consistent except at large mismatches, where performance decreases significantly. For SRE, performance drops dramatically for –ve mismatches but not +ve ones and for SRS performance increases at +ve mismatches. What might cause these individual differences? I believe some of our work investigating the effect of mismatch on sound image fusion and binaural sensitivity might shed some insight into this matter.

What is the effect of mismatch on sound image fusion? Base Apex Electrode # 2 4 6 8 10 12 14 16 18 20 22 Left Electrode # 2 4 6 8 10 12 14 16 18 20 22 Right We conducted a study that examined the effect of interaural mismatch on sound image fusion in CI listeners. 20000 Hz 20 Hz

Find a matched pair via pitch matching Base Apex Electrode # 2 4 6 8 10 12 14 16 18 20 22 Left Electrode # 2 4 6 8 10 12 14 16 18 20 22 Right Δ=0 In order to set up this experiment an interaural pitch matched pair in the center of the electrode array was found through pitch-matching. This is a common technique and assumes that electrodes that have the same pitch excite the same area of the cochlea at the 2 ears. We found a pitch-matched pair near the middle of the array for our experiment. --- In order to set up this experiment an interaural pitch matched pair in the center of the electrode array was found using two tasks. First a pitch rating task was used where the subject was asked to rate the pitch of the sound presented on one of the electrodes using a number from 1 to 100 where 1 was for a low sound and 100 was for a high pitched sound. Here is the result for 1 subject (press). The electrodes that sounds were presented on are shown on the x-axis and the range of ratings on the y-axis. The markers show the average rating over 10 trials and the error bars the standard deviation. On average there is a trend but there is also high variability. We use this task as a guide for picking electrodes to test in the 2nd task. In the second task, subjects compared the pitch of 2 sounds across the ears. Subjects were asked to respond by saying whether the second sound was “much lower”, “lower”, “the same”, “higher” or “much higher” in pitch than the first. We kept one electrode fixed (here, L12) and varied the electrode we stimulated on the other side. We looked for the pair where there was the highest number of same responses (here L12,R11). 20000 Hz 20 Hz

Mismatch conditions Δ=0 Δ=8 Δ=4 Δ=2 Δ=-2 Δ=-4 Δ=-8 1.5 mm Left Right Base Apex Electrode # 2 4 6 8 10 12 14 16 18 20 22 Left Electrode # 2 4 6 8 10 12 14 16 18 20 22 Right Δ=0 We then held one of the electrodes in the pair constant and varied the electrodes we stimulated in the contralateral ear to create mismatched pairs. For this talk I am going to use delta to describe the amount of mismatch in terms of number of electrodes away from the pitch-matched pair. For example, delta = 2 means there was a mismatch of 2 electrode spaces or 3mm. +ve deltas mean that the electrode in the right ear is at a higher frequency place compared to the left ear electrode and –ve deltas mean the left ear electrode is at a higher frequency place compared to the right. 20000 Hz 20 Hz Δ=8 Δ=4 Δ=2 Δ=-2 Δ=-4 Δ=-8 Right higher in frequency Left higher in frequency

What do you hear? For the experiment, subjects were asked to describe what they heard in a 10 alternative, forced choice task. This is the GUI for the experiment. The 10 categories were divided into 3 groups: Top row – single fused image Middle row – split image Bottom row – diffuse sound image 9 post-lingually deafened, Cochlear Nucleus implant users The stimuli were 300-ms, 100-pulse per second, constant amplitude pulse trains delivered using synchronized research processors directly to the implants. No ITD/ILD applied 20 trials per mismatch condition CI = 140 trials

Mismatch affects where a sound is perceived Orientation: Icons are along the y-axis, mismatch is along x. Δ=0 is the pitch-matched pair. Color represents the percentage of trials at a particular mismatch a particular category was chosen. Things to note: Single fused image Centered at Δ=0 Lateralizes with mismatch (here IBD follows LOW freqencies) Left higher in frequency Right higher in frequency

Mismatch affects people differently Things to note: Subject follows high frequencies Split image at large mismatch Left higher in frequency Right higher in frequency Left higher in frequency Right higher in frequency

Increasing mismatch leads to more split sound images If I pool the group data and count the number of trials where multiple sound sources were perceived, we can see here that with increasing mismatch, there is an increasing percentage of trials where multiple sound sources are perceived.

Most people lateralize towards high frequencies Here I’ve plotted the proportion of subjects who are “high frequency followers”, “low frequency followers” and “multiple image perceivers” and we can see that the proportions are not the same. About 2/3 of subjects fuse and lateralize towards the side of the higher frequency electrode.

Possible Implications Target Target Target What are the possible implications? Well if we have a target speaker in front, A small mismatch would lead to a lateralized sound image (which would happen for most listeners); or At a large mismatch, the target talker may become a split image Lateralized sound image Split sound image

How does mismatch affect sounds off the mid-line? Typical values: ILD = 0; ±2; ±5; ±10 CU ITD = 0; ±100; ±200; ±400; ±800 µs Negative values mean sound is to the left. Now what about a sound located off the mid-line? We take the same stimuli and now apply either an ILD or ITD. ILDs were applied by changing the current levels applied at each ear which changed the loudness of the stimuli at one ear relative to the other, and ITDs were applied by delaying the onset of one sound relative to the other. Subjects were asked to indicate on the GUI how many sound they heard and where they heard each sound. For each sound perceived, a colored bar appeared on the face. Here, we have an example of two sounds. The subject was also asked to rank the dominance of each sound, where the location of the primary or “most dominant” sound was placed on the top bar, etc.

Good lateralization at no mismatch Here’s data for the same 2 subjects as before. We see that with no mismatch, these subjects hear the sound lateralize from the left to the right with changes in ILD.

Mismatch affect ILDs Now with large mismatches, we see that the lateralization curves shift, such that 0 ILD is no longer heard at the center of the head. It would appear that subjects can still lateralize sounds since loudness seems to compensate for the shift.

Mismatch affect ILDs If I show the intermediate mismatches, we see that there is a systematic shift of the lateralization curves.

Mismatch affect ITDs even more However, with ITDs, subjects are unable to lateralize when there is a large mismatch in the place of stimulation across the ears.

Non-midline crossings increase with mismatch So how many subjects are unable to lateralize sounds with increasing mismatches? Here, I’ve plotted the percentage of subjects whose lateralization curves do not cross the midline at different amounts of mismatch. We observe that with increasing mismatch, there is an increasing number of subjects who do not have lateralization curves that cross the midline and that mismatch affects ITD lateralization more than ILD lateralization

Possible Implications Masker Masker Masker What are the possible implications? Well if we have a masker speaker on the side, with mismatch They may be perceived on the wrong side; or They too may become a split image Lateralized sound image Split sound image

Possible implications on SRM Target Maskers Lateralized Image Target So, what are the possible implications for SRM? Mismatch might cause both Target and Masker to lateralize to the same side => No SRM Masker

Possible implications on SRM Target Maskers Lateralized Image Target Maskers Lateralized Image Target 2. Target and Maskers might lateralize to different sides => possibly increasing SRM Masker

Possible implications on SRM Target Maskers Lateralized Image Target Maskers Lateralized Image Target 3. Target and Maskers are both split => zero and negative SRM due to confusion Target Maskers Split Image Masker

Summary Interaural place-of-stimulation mismatch leads to lateralized or split sound images which affects perceived location of sounds Depending on location of target and masker, interaural mismatch might increase or hinder spatial release from masking

Thank you We’d like to thank our research participants for participating in our experiments, and Cochlear Ltd. for supplying the equipment and technical support. This work is supported by NIH-NIDCD R01003083 (Litovsky) and in part by a core grant to the Waisman Center from the NICHD (P30 HD03352).