Hearing Aids and Hearing Impairments Part II Meena Ramani 02/23/05.

Slides:



Advertisements
Similar presentations
Revised estimates of human cochlear tuning from otoacoustic and behavioral measurements Christopher A. Shera, John J. Guinan, Jr., and Andrew J. Oxenham.
Advertisements

SOUND PRESSURE, POWER AND LOUDNESS MUSICAL ACOUSTICS Science of Sound Chapter 6.
Figures for Chapter 7 Advanced signal processing Dillon (2001) Hearing Aids.
Introduction to MP3 and psychoacoustics Material from website by Mark S. Drew
Pulse Techniques. Off-Resonance Effects Initial magnetization along z x-pulse (  = 0) On-resonance: M z -> -M y Off-resonance: phase 
Stefan Bleeck, Institute of Sound and Vibration Research, Hearing and Balance Centre University of Southampton.
Room Acoustics: implications for speech reception and perception by hearing aid and cochlear implant users 2003 Arthur Boothroyd, Ph.D. Distinguished.
Hearing and Deafness 2. Ear as a frequency analyzer Chris Darwin.
CS 551/651: Structure of Spoken Language Lecture 11: Overview of Sound Perception, Part II John-Paul Hosom Fall 2010.
Hearing and Deafness Outer, middle and inner ear.
Speech Science XII Speech Perception (acoustic cues) Version
Page 0 of 34 MBE Vocoder. Page 1 of 34 Outline Introduction to vocoders MBE vocoder –MBE Parameters –Parameter estimation –Analysis and synthesis algorithm.
Speaking Style Conversion Dr. Elizabeth Godoy Speech Processing Guest Lecture December 11, 2012.
Electroacoustic Testing of DSP Hearing Aids Christine Cameron & Mary Hostler MCHAS Team University of Manchester.
A.Diederich– International University Bremen – Sensation and Perception – Fall Frequency Analysis in the Cochlea and Auditory Nerve cont'd The Perception.
Interrupted speech perception Su-Hyun Jin, Ph.D. University of Texas & Peggy B. Nelson, Ph.D. University of Minnesota.
Sound Transmission and Echolocation Sound transmission –Sound properties –Attenuation Echolocation –Decoding information from echos.
Fitting Formulas Estimate amplification requirements of individual patients Maximize intelligibility of speech Provide good overall sound quality Keep.
Spectral centroid 6 harmonics: f0 = 100Hz E.g. 1: Amplitudes: 6; 5.75; 4; 3.2; 2; 1 [(100*6)+(200*5.75)+(300*4)+(400*3.2)+(500*2 )+(600*1)] / = 265.6Hz.
Relationship between perception of spectral ripple and speech recognition in cochlear implant and vocoder listeners L.M. Litvak, A.J. Spahr, A.A. Saoji,
Sound source segregation (determination)
1 Recent development in hearing aid technology Lena L N Wong Division of Speech & Hearing Sciences University of Hong Kong.
1 Audio Compression Multimedia Systems (Module 4 Lesson 4) Summary: r Simple Audio Compression: m Lossy: Prediction based r Psychoacoustic Model r MPEG.
Normalization of the Speech Modulation Spectra for Robust Speech Recognition Xiong Xiao, Eng Siong Chng, and Haizhou Li Wen-Yi Chu Department of Computer.
Figures for Chapter 6 Compression
Hearing Aids and Hearing Impairments
SIGNAL PROCESSING IN HEARING AIDS
Speech Signal Processing
CSD 5400 REHABILITATION PROCEDURES FOR THE HARD OF HEARING Auditory Perception of Speech and the Consequences of Hearing Loss.
1 SPEECH PROCESSING FOR BINAURAL HEARING AIDS Dr P. C. Pandey EE Dept., IIT Bombay Feb’03.
Second International Conference on Intelligent Interactive Technologies and Multimedia (IITM 2013), March 2013, Allahabad, India 09 March 2013 Speech.
Speech Enhancement Using Spectral Subtraction
Page 0 of 23 MELP Vocoders Nima Moghadam SN#: Saeed Nari SN#: Supervisor Dr. Saameti April 2005 Sharif University of Technology.
Speech Science Fall 2009 Oct 28, Outline Acoustical characteristics of Nasal Speech Sounds Stop Consonants Fricatives Affricates.
Department of Electrical & Computer Engineering Auditory Perception Meena Ramani 04/09/2004.
Chapter 5: Normal Hearing. Objectives (1) Define threshold and minimum auditory sensitivity The normal hearing range for humans Define minimum audible.
Studies of Information Coding in the Auditory Nerve Laurel H. Carney Syracuse University Institute for Sensory Research Departments of Biomedical & Chemical.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
Need for cortical evoked potentials Assessment and determination of amplification benefit in actual hearing aid users is an issue that continues to be.
Applied Psychoacoustics Lecture 3: Masking Jonas Braasch.
SOUND PRESSURE, POWER AND LOUDNESS MUSICAL ACOUSTICS Science of Sound Chapter 6.
IIT Bombay {pcpandey,   Intro. Proc. Schemes Evaluation Results Conclusion Intro. Proc. Schemes Evaluation Results Conclusion.
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
Laboratory for Experimental ORL K.U.Leuven, Belgium Dept. of Electrotechn. Eng. ESAT/SISTA K.U.Leuven, Belgium Combining noise reduction and binaural cue.
Active Microphone with Parabolic Reflection Board for Estimation of Sound Source Direction Tetsuya Takiguchi, Ryoichi Takashima and Yasuo Ariki Organization.
The Ear As a Frequency Analyzer Reinier Plomp, 1976.
Microphone Array Project ECE5525 – Speech Processing Robert Villmow 12/11/03.
Introduction to psycho-acoustics: Some basic auditory attributes For audio demonstrations, click on any loudspeaker icons you see....
EE Dept., IIT Bombay Workshop “Radar and Sonar Signal Processing,” NSTL Visakhapatnam, Aug 2015 Coordinator: Ms. M. Vijaya.
Speech Perception.
IIT Bombay 17 th National Conference on Communications, Jan. 2011, Bangalore, India Sp Pr. 1, P3 1/21 Detection of Burst Onset Landmarks in Speech.
Fletcher’s band-widening experiment (1940)
Motorola presents in collaboration with CNEL Introduction  Motivation: The limitation of traditional narrowband transmission channel  Advantage: Phone.
SOUND PRESSURE, POWER AND LOUDNESS
SPATIAL HEARING Ability to locate the direction of a sound. Ability to locate the direction of a sound. Localization: In free field Localization: In free.
Speech and Singing Voice Enhancement via DNN
Speech Enhancement Summer 2009
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
PSYCHOACOUSTICS A branch of psychophysics
Precedence-based speech segregation in a virtual auditory environment
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
Loudness level (phon) An equal-loudness contour is a measure of sound pressure (dB SPL), over the frequency spectrum, for which a listener perceives a.
Liverpool Keele Contribution.
A Smartphone App-Based
Speech Perception.
CHAPTER 10 Auditory Sensitivity.
Speech Perception (acoustic cues)
Auditory, Tactical, and Olfactory Displays
Presenter: Shih-Hsiang(士翔)
COPYRIGHT © All rights reserved by Sound acoustics Germany
Presentation transcript:

Hearing Aids and Hearing Impairments Part II Meena Ramani 02/23/05

Discussion Time!

Summarize  Facts on Hearing Loss  Hearing Aids  Cochlea-IHC and OHC  Presbycusis  Decreased Audibility  Decreased Frequency Resolution  Decreased Temporal resolution  Decreased Dynamic Range  Amplification Techniques  Linear  Compressive- Single/MultiBand Room for improvement Huge Market BTE,ITE,ITC,CIC OHCs: Sharpen the traveling wave Provide an amplification for soft sounds(40-50 dB SPL) HL in aging ears Occurs due to damage to OHCs 1) HA has to provide more gain at HFs. 2) HAs less gain at LFs Noise Removal 3) Fast acting compression 4) Compressive Amplification Linear-too much gain Compressive- Overshoots/undershoots Multiband vs Singleband

Outline Temporal Resolution Frequency Resolution Noise Reduction Techniques Conclusion

Temporal Resolution What is temporal resolution? What happens to temporal resolution for the HI? What does poor temporal resolution result in? Implications for HA design

What is temporal resolution? Speech has a lot of temporal information like the presence or absence of acoustic excitation, the periodicity or aperiodicity of excitation _______, ________,______  Speech Envelope Slowly varying Carries Information: Consonants, voicing, phoneme boundaries, syllable boundaries, stress etc. Lip reading and speech envelope  Modulation Perception Changing the depth of modulation of the envelope Noise and Reverberations  Gap detection threshold Psychoacoustic measure For normals: 2.5ms Relationship between gap detection thresholds and SRTs in noise (Consonant recognition requires temporal structures)

Temporal resolution for the HI Experimental setup for Modulation Perception :  TMTF- Temporal Modulation Transfer Function  Sinusoidal modulation of broadband noise  Modulation detection threshold in terms of freq.  Comparison with normals: threshold shift /SL Results :  Poor modulation perception is because of reduced listening bandwidth  Same behavior at SL inputs for normals and HI Results for Gap detection measures :  Normals: GDT reduces as the frequency of the noise bands increases  Same behavior at SL inputs for normals and HI Signals which are made audible to HI have same temporal resolution as the normals

Temporal resolution for the HI (contd.) Difference in loudness levels between the envelope maxima and minima is > for impaired ear than normal This leads one to assume that impaired ear will perceive modulation depth changes better perceptually/Louder Circles->equal modulation strength. Contradicts the TMTF results?  JND vs Perception  Noise also enhanced

Temporal resolution for the HI (contd.) Effect of Compression on Modulation Use compression to provide loudness correction. Fast acting/Syllabic compression 3:1 compression Modulation depth (dB)=20logm AM factor: (1+msin(w m t)) Reduces the modulation depth by~9.5dB

Implications for HA design Syllabic compression can compensate for abnormal sensitivity to AM This compression also improves the discrimination of envelopes having a DR>10dB But reducing the spectral cues causes in low SNR conditions a low SI.

Frequency Resolution What is Frequency resolution? What happens to freq. resolution for the HI? What does poor freq. resolution result in? Implications for HA design

What is Frequency resolution? If change in spectrum of speech causes some change in shape of excitation along basilar membrane => change exceeds listeners frequency resolution Else => frequency resolution was not fine enough to discriminate the spectral changes

Frequency resolution for HI Statement: Cochlear damage results in poor freq. resolution But Auditory filter bandwidth increases with stimulus level… HOW DO YOU MEASURE FREQ. RESOLUTION? Experimental setup:  Need normals and the HI to be at the same sensation level(SL)  Normals: Add broad band background noise to elevate threshold Results:  Freq. resolution measured via tuning curves was worse for HI  More Upward spread of masking since LF slope of filter is shallower than for normals. Conclusion: Freq. resolution is impaired by both:  Damaged auditory system  Necessity to listen to high stimulus

What does poor frequency resolution result in?  Loose Spectral Cues Formant peak information is lost Smooths the internal spectral contrasts  Inability to distinguish between vowels F1 & F2 frequencies  Important cue for vowel ID  Increase in upper spread of masking  CVR-Consonant Vowel Recognition HI have more problems understanding the speech in noise when compared to normals

Implications for HA design Fact: For the HI, we have broader auditory filters  Sharpen spectral contrast Narrow the BW of spectral peaks Decrease level of spectral valleys  Not too much success since the broad filters overwhelm the sharpening technique Multiband/wideband design:  Reduction in spectral cues  For Multiband correlate AGC in each band

Noise Reduction HI people have abnormal difficult understanding speech in noise.

Noise Reduction HI need an SNR of 9dB  Broader auditory filters, reduced suppressions  Upper spread of masking SNR improvement doesn’t often correlate with improved SI! Noise removal algorithms  Single-microphone techniques  Multi-microphone techniques

Single-microphone techniques General Considerations Single stream has speech+noise Need to evaluate continuously which frames have speech and which have noise Improvements in SNR do not relate directly to improvements in SI  Need to evaluate performance of algorithm using Listening SI tests

Single-microphone techniques Frequency specific gain reduction BILL-Bass increase at Low Levels  For noise reduction, bass decrease at high levels  Reduces LFs when the average gain in that region is high  Theoretically should help since the HI LFs mask the HF  LF has information about consonant features such as nasality, voicing etc which was lost Cook et al in 1996 showed that if noise is LF, then HPF the speech resulted in significant improvement of SI Festen et al in 1990 Envelope minima technique, reduce gain per band so that envelope minima (noise) is closer to hearing threshold level Dynamic range based technique: attenuation for noise band inversely proportional to the measured DR

Single-microphone techniques Spectral subtraction Subtract spectral magnitude of the noise estimate from the short term spectral magnitude of the signal  Assumes stationary noise  Uses same phase for the final noise reduced signal  SNR improves but SI is same. Removes noise like cues required for fricatives.

Multiple microphone techniques What is Array Processing? Omni directional microphones: 15mm separation between any two Low frequency roll off of 6dB/octave Figure A typical configuration for a two- microphone (Mic) directional system. The delay to the back microphone determines the angle of the null in the directional pattern. Figure Two directional patterns typically associated with hearing aid directional microphones. The angle represents the direction from which the sound is approaching the listener, with 0 degrees representing directly in front of the listener. The distance from the origin at a given angle represents the gain applied for sound from arriving from that direction, ranging here from 0 to 25dB. The patterns are a cardioid (left) and a hypercardioid (right).

Beamforming Frequency Dependent Frequency Independent Delay and sum

Comparison with noise suppressor Noise suppressor (NS) is the standard one used on iDEN phones

Noise Cancellation Use LMS algorithm Problem is some speech is fed back to reference mic and gets canceled with noise. Figure A typical two-microphone noise cancellation system. Ideally, the primary microphone measures a mixture of the interfering noise and the target speech, and the reference microphone measures only a transformation of the interfering noise.

Conclusion Parameters selection and fitting is a very difficult problem Algorithms can make the sound more audible but not more intelligible IHC have been ignored so far but they could have a role too. It is difficult to get subjective scores from HI populations No objective method can account for the non- linearities introduced by compression Wearable HAs are an option for research but are inconvenient

Array fundamentals Speaker tracking is not possible with single microphone Multiple microphones facilitate spatiotemporal filtering Setup consists of two microphones with the first microphone assumed as origin Distance of the wavefront from microphone is The direction of source is given by