Stefan Bleeck, Institute of Sound and Vibration Research, Hearing and Balance Centre University of Southampton.

Slides:



Advertisements
Similar presentations
Hearing and Deafness 1. Anatomy & physiology Chris Darwin Web site for lectures, lecture notes and filtering lab:
Advertisements

Revised estimates of human cochlear tuning from otoacoustic and behavioral measurements Christopher A. Shera, John J. Guinan, Jr., and Andrew J. Oxenham.
Advanced Speech Enhancement in Noisy Environments
Hearing and Deafness 2. Ear as a frequency analyzer Chris Darwin.
Hearing and Deafness Outer, middle and inner ear.
MIMICKING THE HUMAN EAR Philipos Loizou (author) Oliver Johnson (me)
Pitch organisation in Western tonal music. Pitch in two dimensions Pitch perception in music is often thought of in two dimensions, pitch height and pitch.
Hearing Aka: Audition. Frequency the number of complete wavelengths that pass through point at a given time. This determines the pitch of a sound.
M.Sc. in Medical Engineering
A.Diederich– International University Bremen – Sensation and Perception – Fall Frequency Analysis in the Cochlea and Auditory Nerve cont'd The Perception.
Physiology of the cochlea Mechanical response of cochlea in response to sound Two major functions: 1. Analysis of sound into components: Frequency/Spectral.
Hearing and Deafness 1. Anatomy & physiology Chris Darwin Web site for lectures, lecture notes and filtering lab:
Structure and function
Relationship between perception of spectral ripple and speech recognition in cochlear implant and vocoder listeners L.M. Litvak, A.J. Spahr, A.A. Saoji,
1 New Technique for Improving Speech Intelligibility for the Hearing Impaired Miriam Furst-Yust School of Electrical Engineering Tel Aviv University.
Chapter 6: The Human Ear and Voice
Unit 4: Sensation & Perception
ENGINEERING BIOMEDICAL Michael G. Heinz Background: PhD at MIT –Speech and Hearing Bioscience and Technology Postdoctoral Fellow at Johns Hopkins Univ.
Sensation and Perception: Hearing
Hearing.
Chapter 4 Powerpoint: Hearing
HEARING. Audition  What is Audition?  Hearing  What sounds do we hear the best?  Sounds with the frequencies in the range corresponding to the human.
AIM: How do we hear?. Opponent Process Theory Hering proposed that we process four primary colors combined in pairs of red-green, blue- yellow, and black-white.
DO NOW: Put your homework packet together and get your reading notes out to be checked. THEN answer: Explain the Young-Helmholtz trichromatic theory.
CAN YOU HEAR ME NOW? Hearing. What to Expect/Objectives  Describe what hearing is  Describe the pressure waves that experiences as sound  Describe.
Hearing Impairment Hair cells are responsible for translating mechanical information into neural information. Thus, with damaged hair cells, the auditory.
Humans can hear sounds at frequencies from about 20Hz to 20,000Hz.
From Auditory Masking to Supervised Separation: A Tale of Improving Intelligibility of Noisy Speech for Hearing- impaired Listeners DeLiang Wang Perception.
1 Hearing or Audition Module 14. Hearing Our auditory sense.
Sensation Vision The Eye Theories Hearing The Ear Theories Other Senses Smell Taste Pain Gestalt Principles Perceptual Constancies Perception Basic Principles.
Hearing Our auditory sense We hear sound WAVES Frequency: the number of complete wavelengths that pass through point at a given time. This determines.
1 PSYCHOLOGY (8th Edition, in Modules) David Myers PowerPoint Slides Aneeq Ahmad Henderson State University Worth Publishers, © 2007.
IB Assessment Statements Option I-1, The Ear and Hearing: I.1.1.Describe the basic structure of the human ear. I.1.2.State and explain how sound pressure.
Gammachirp Auditory Filter
A NEW FEATURE EXTRACTION MOTIVATED BY HUMAN EAR Amin Fazel Sharif University of Technology Hossein Sameti, S. K. Ghiathi February 2005.
Sensation and Perception: Hearing. Sound Sound comes in waves. The waves are produced by vibration. Sound is vibration Ex. Clap your hands causes the.
IIT Bombay {pcpandey,   Intro. Proc. Schemes Evaluation Results Conclusion Intro. Proc. Schemes Evaluation Results Conclusion.
Humans can hear sounds at frequencies from about 20Hz to 20,000Hz.
You better be listening… Auditory Senses Sound Waves Amplitude  Height of wave  Determines how loud Wavelength  Determines pitch  Peak to peak High.
Chapter 4 Sensation and Perception. The Ear Audition = hearing Audition = hearing Sounds = mechanical energy typically caused by vibrating objects Sounds.
HEARING Module 20. Hearing – sound waves  Audition – the sense or act of hearing  Frequency – the number of complete wavelengths that pass a point in.
What can we expect of cochlear implants for listening to speech in noisy environments? Andrew Faulkner: UCL Speech Hearing and Phonetic Sciences.
AP Psychology Unit 4 Module 20
대학원 생체신호처리 - 4 이상민.
Auditory System…What??? It plays an important role in language development and social interactions… Plus…it alerts us to dangerous situations! The auditory.
Hearing Module 14.
Oregon Health & Science University
Speech Enhancement Summer 2009
Precedence-based speech segregation in a virtual auditory environment
Review: Hearing.
Perceptual Constancies
Audition (Hearing).
Presenter: Prof.Dr.-Eng. Gheorghe-Daniel Andreescu
Peripheral Auditory System
A proposed Proof-of-Concept experiment
Hearing Aka: Audition.
SENSORY PHYSIOLOGY: THE EAR
Hearing, not trying out for a play
Hearing: The Nature of Sound
Perceptual Constancies
Hearing AKA: Audition.
Chapter 5: Sensation Hearing.
Peripheral Auditory System
Sensation Notes 5-3 (obj 11-16)
The Human Ear.
Fang Du, Dr. Christina L. Runge, Dr. Yi Hu. April 21, 2018
How We Hear.
Hearing Aka: Audition.
Detection of Cochlear Amplification and Its Activation
L6 – Hearing and the Ear Learning Objectives:
Presentation transcript:

Stefan Bleeck, Institute of Sound and Vibration Research, Hearing and Balance Centre University of Southampton

 Can sparse coding help to overcome problems caused by hearing loss? ◦ overview of the hearing process ◦ Examples of sparse algorithms for hearing aids and cochlear implants ◦ Preliminary results 5/11/2012

5/11/2012

5/11/2012 Important for sound localization, linear => boring

5/11/2012 Important to explain limits of hearing, linear => boring

5/11/2012  Contained within bony labyrinth in temporal bone  Cochlea does hearing  Semicircular canals+utricle does balance  Same mechanism, nerve, evolution, similar problems

5/11/ frequency mapping

5/11/  Stereocilia detect vibrations within cochlea.  Introduce half-wave rectification  Nonlinear

13 orders of magnitude 10 Power (watts/m 2 )

5/11/ ABSOLUTE THRESHOLD CURVE membrane moves m Threshold as function of Frequency

5/11/2012  Amplitude (nonlinear amplification)  Frequencies (combination tones)  compression Demo: sweeps

12 Distance along BM (mm) BM displacement (nm) damaged “passive” healthy “active” OHCs inject energy in this region OHCs provide up to 40 dB amplification (= factor of 100) Travelling Wave Envelope on Basilar Membrane due to Pure Tone Stimulus:

 50% of 60 year old, 90% of 80 year old  Hearing aids are not good enough  ‘damage’ €2.4 Billion per year in EU  Lack of research funding today 5/11/2012

5/11/2012 Electron micrographs of cochlear hair cells. Left: healthy, right: damaged by noise exposure.

5/11/2012 Hearing impairment loosing audibility, Also widening of filter both results in difficulties to understand language, especially in noise

Listening in noise 0 dB 40 dB-15 dB ASR Normal Hearing Impaired SNR Word recognition 100% 0% Aided Un-aided 50%

5/11/2012

5/11/2012

5/11/2012 Problem: Hearing loss constitutes a bottle neck: not all information can get through Solution: extract less, but important information -Extract content based on Information not on Energy -Specifically speech related information

2 Neural representation: (Transformation) 3 Denoising (sparsification) 1 periphery model

5/11/2012 ‘Sparse’ algorithms developed in our group noise

5/11/2012 Filter bank

 Non-negative matrix factorization ◦ Matrix Z is factorised into two non-negative matrices W and H (basis vectors (5) and activity over time) ◦ (motivated by the processing in CI and auditory neurons) ◦ Z here is the ‘envelopegram’ (22 channels, 128 pt) ◦ Factorization using Euclidean cost function: ◦ Sparseness constrained: 5/11/2012 g(H)= regularity function λ= sparsity factor

Iterative algorithm to minimize the cost function by gradient decent: λ depends on SNR because of trade-off intelligibility - quality low noise: no sparsification high noise: lots Task: fine out how! Online experiment (restricted by speed of hardware) Offline experiment (unrestricted) 5/11/2012

5/11/2012 For ‘bin’, ‘pin’, ‘din’, ‘tin’ Z W H

5/11/2012

5/11/2012 On-line experimental set up:

 22 channel filter bank  16 ms frames  Gaussian noise SNR=5 dB clean noisy denoised time frequency

5/11/2012 Results from CI listeners in online experiment (problems with iteration!)

5/11/2012 Results from CI listeners in offline experiment results for all participants Averaged Best sparsification as function of snr:

Conclusions: Sparse coding can help reduce acoustic information in a useful way Development still in its infancy, hardware restrictions still relevant High impact research field with lots of potential funding Strength of our group: clinical evaluation, weakness at the moment: lack of signal processing experts

5/11/2012 Hu, H., Li, G., Chen, L., Sang, J., Wang, S., Lutman, M. E., & Bleeck, S. (2011). Enhanced sparse speech coding strategy for cochlear implants. European Signal Processing Conference (EUSIPCO). Hu, H., Taghia, J., Sang, J., Taghia, J., Mohammadiha, N., Azarpour, M., Dokku, R., et al. (2011). Speech Enhancement via Combination of Wiener Filter and Blind Source Separation. International Conference on Intelligent Systems and Knowledge Engineering. Sang, J., Hu, H., Li, G., Lutman, M. E., & Bleeck, S. (2011a). Application of a sparse coding strategy to enhance speech perception for hearing aid users. British Society of Audiology Short Papers Meeting. Sang, J., Hu, H., Li, G., Lutman, M. E., & Bleeck, S. (2011b). Enhanced Sparse Speech Processing Strategy in Cochlear Implants. Conference on implantable Auditory Prostheses (CIAP). Sang, J., Li, G., Hu, H., Lutman, M. E., & Bleeck, S. (2011a). Supervised Sparse Coding in Cochlear Implants. Conference on implantable Auditory Prostheses (CIAP). Sang, J., Li, G., Hu, H., Lutman, M. E., & Bleeck, S. (2011b). Supervised Sparse Coding Strategy in Hearing Aids. Annual Conference of the International Speech Communication Association (INTERSPEECH). Bleeck, S., Wright, M. C. M., & Winter, I. M. (2012). Speech enhancement inspired by auditory modelling. International Symposium on Hearing. Hu, H., Mohammadiha, N., Taghia, J., Leijon, A., Lutman, M. E., Bleeck, S., & Wang, S. (2012). Sparsity Level in a Non- negative Matrix Factorization Based Speech Strategy in Cochlear Implants. EUSIPCO. Li, G, Lutman, M. E., Wang, S., & Bleeck, S. (2012). Relationship between speech recognition in noise and sparseness. International Journal of Audiology, 51(2), 75–82. doi: / Sang, J., Hu, H., Zheng, C., Li, G., Lutman, M. E., & Bleeck, S. (2012). Evaluation of a Sparse Coding Shrinkage Algorithm in Normal Hearing and Hearing Impaired Listeners. EUSIPCO (pp. 1–5).