Auditory Localization in Rooms: Acoustic Analysis and Behavior

Slides:



Advertisements
Similar presentations
Auditory Localisation
Advertisements

Evaluation of Dummy-Head HRTFs in the Horizontal Plane based on the Peak-Valley Structure in One-degree Spatial Resolution Wersényi György SZÉCHENYI ISTVÁN.
1 3D sound reproduction with “OPSODIS” and its commercial applications Takashi Takeuchi, PhD Chief Technical Officer OPSODIS Limited Institute of Sound.
Binaural Hearing Or now hear this! Upcoming Talk: Isabelle Peretz Musical & Non-musical Brains Nov. 12 noon + Lunch Rm 2068B South Building.
Spatial Perception of Audio J. D. (jj) Johnston Neural Audio Corporation.
Localizing Sounds. When we perceive a sound, we often simultaneously perceive the location of that sound. Even new born infants orient their eyes toward.
Improvement of Audibility for Multi Speakers with the Head Related Transfer Function Takanori Nishino †, Kazuhiro Uchida, Naoya Inoue, Kazuya Takeda and.
Source Localization in Complex Listening Situations: Selection of Binaural Cues Based on Interaural Coherence Christof Faller Mobile Terminals Division,
3-D Sound and Spatial Audio MUS_TECH 348. Cathedral / Concert Hall / Theater Sound Altar / Stage / Screen Spiritual / Emotional World Subjective Music.
Visually-induced auditory spatial adaptation in monkeys and humans Norbert Kopčo Center for Cognitive Neuroscience, Duke University Hearing Research Center,
Development of sound localization
ICA Madrid 9/7/ Simulating distance cues in virtual reverberant environments Norbert Kopčo 1, Scott Santarelli, Virginia Best, and Barbara Shinn-Cunningham.
Effect of roving on spatial release from masking for amplitude-modulated noise stimuli Norbert Kopčo *, Jaclyn J. Jacobson, and Barbara Shinn-Cunningham.
Hearing & Deafness (3) Auditory Localisation
AUDITORY PERCEPTION Pitch Perception Localization Auditory Scene Analysis.
Spectral centroid 6 harmonics: f0 = 100Hz E.g. 1: Amplitudes: 6; 5.75; 4; 3.2; 2; 1 [(100*6)+(200*5.75)+(300*4)+(400*3.2)+(500*2 )+(600*1)] / = 265.6Hz.
Sound source segregation (determination)
Acoustical Society of America, Chicago 7 June 2001 Effect of Reverberation on Spatial Unmasking for Nearby Speech Sources Barbara Shinn-Cunningham, Lisa.
Alan Kan, Corey Stoelb, Matthew Goupell, Ruth Litovsky
Improved 3D Sound Delivered to Headphones Using Wavelets By Ozlem KALINLI EE-Systems University of Southern California December 4, 2003.
SOUND IN THE WORLD AROUND US. OVERVIEW OF QUESTIONS What makes it possible to tell where a sound is coming from in space? When we are listening to a number.
LINKWITZ LAB What are the On-axis & Off-axis
Applied Psychoacoustics Lecture: Binaural Hearing Jonas Braasch Jens Blauert.
Supervisor: Dr. Boaz Rafaely Student: Limor Eger Dept. of Electrical and Computer Engineering, Ben-Gurion University Goal Directional analysis of sound.
Chapter 12: Auditory Localization and Organization
Chapter 5: Normal Hearing. Objectives (1) Define threshold and minimum auditory sensitivity The normal hearing range for humans Define minimum audible.
Studies of Information Coding in the Auditory Nerve Laurel H. Carney Syracuse University Institute for Sensory Research Departments of Biomedical & Chemical.
Calibration of Consonant Perception in Room Reverberation K. Ueno (Institute of Industrial Science, Univ. of Tokyo) N. Kopčo and B. G. Shinn-Cunningham.
Developing a model to explain and stimulate the perception of sounds in three dimensions David Kraljevich and Chris Dove.
 Space… the sonic frontier. Perception of Direction  Spatial/Binaural Localization  Capability of the two ears to localize a sound source within an.
Audio Systems Survey of Methods for Modelling Sound Propagation in Interactive Virtual Environments Ben Tagger Andriana Machaira.
‘Missing Data’ speech recognition in reverberant conditions using binaural interaction Sue Harding, Jon Barker and Guy J. Brown Speech and Hearing Research.
3-D Sound and Spatial Audio MUS_TECH 348. Physical Modeling Problem: Can we model the physical acoustics of the directional hearing system and thereby.
L INKWITZ L AB S e n s i b l e R e p r o d u c t i o n & R e c o r d i n g o f A u d i t o r y S c e n e s Hearing Spatial Detail in Stereo Recordings.
Jens Blauert, Bochum Binaural Hearing and Human Sound Localization.
Spatial and Spectral Properties of the Dummy-Head During Measurements in the Head-Shadow Area based on HRTF Evaluation Wersényi György SZÉCHENYI ISTVÁN.
Hearing Research Center
This research was supported by Delphi Automotive Systems
Human Detection and Localization of Sounds in Complex Environments W.M. Hartmann Physics - Astronomy Michigan State University QRTV, UN/ECE/WP-29 Washington,
Auditory Neuroscience 1 Spatial Hearing Systems Biology Doctoral Training Program Physiology course Prof. Jan Schnupp HowYourBrainWorks.net.
Listeners weighting of cues for lateral angle: The duplex theory of sound localization revisited E. A. MacPherson & J. C. Middlebrooks (2002) HST. 723.
IIT Bombay 17 th National Conference on Communications, Jan. 2011, Bangalore, India Sp Pr. 1, P3 1/21 Detection of Burst Onset Landmarks in Speech.
Hearing in Distance Or Where is that sound? Today: Isabelle Peretz Musical & Non-musical Brains Nov. 12 noon + Lunch Rm 2068B South Building.
3-D Sound and Spatial Audio MUS_TECH 348. Are IID and ITD sufficient for localization? No, consider the “Cone of Confusion”
Energy, Stereoscopic Depth, and Correlations. Molecules Levels of Information Processing in the Nervous System 0.01  m Synapses 1m1m Neurons 100 
Development of Sound Localization 2 How do the neural mechanisms subserving sound localization develop?
PSYC Auditory Science Spatial Hearing Chris Plack.
3-D Sound and Spatial Audio MUS_TECH 348. Environmental Acoustics, Perception and Audio Processing: Envelopment.
Fletcher’s band-widening experiment (1940)
The role of reverberation in release from masking due to spatial separation of sources for speech identification Gerald Kidd, Jr. et al. Acta Acustica.
SPATIAL HEARING Ability to locate the direction of a sound. Ability to locate the direction of a sound. Localization: In free field Localization: In free.
Fundamentals of Sensation and Perception
3-D Sound and Spatial Audio MUS_TECH 348. What do these terms mean? Both terms are very general. “3-D sound” usually implies the perception of point sources.
Emotional response to sound Influence of spatial determinants Master’s Thesis in the Master’s programme in Sound and Vibration Fredrik Hagman Presented.
Sound Localization and Binaural Hearing
Precedence-based speech segregation in a virtual auditory environment
A B C D VIS AUD V azi ele stim resp gaze shift error Figure 12-1.
The barn owl (Tyto alba)
3) determine motion and sound perceptions.
1.
Auditorium acoustic (continued)
Volume 62, Issue 1, Pages (April 2009)
RESULTS: Individual Data
Hearing Spatial Detail
Visually-induced auditory spatial adaptation in monkeys and humans
Volume 62, Issue 1, Pages (April 2009)
Localizing Sounds.
Auditory perception: The near and far of sound localization
3 primary cues for auditory localization: Interaural time difference (ITD) Interaural intensity difference Directional transfer function.
Week 13: Neurobiology of Hearing Part 2
Presentation transcript:

Auditory Localization in Rooms: Acoustic Analysis and Behavior Norbert Kopčo1 Barbara Shinn-Cunningham1,2 Boston University Hearing Research Center 1Cognitive and Neural Systems, Boston University 2Biomedical Engineering, Boston University

Localization in a room: Degraded directional hearing Improved distance perception Goal: Analyze how localization cues are influenced by listener position in a room source azimuth and distance around listener Compare predictions to behavior Outline: Effect of Room on Spatial Cues: Spectral Cues, Interaural Level Differences, Interaural Time Differences Human Localization Performance Conclusions and Future work

Methods Measure KEMAR HRTFs 4 room positions, azimuths 0 – 90° (15° steps), dist 15, 40, 100 cm Analyze effect of reverberation on ILD, ITD, spectrum In terms of mean value and variability Human performance mean and std dev in perceived azimuth within 0-90°, 0.15-1m Positions of KEMAR and listeners Positions of sound source Room 5x9 meters, T60 ~700 ms center wall corner back

Reverberation Alters Spectral Cues HRTF analysis Reverberation Alters Spectral Cues Anechoic Reverberant Left Ear Normalized Amplitude Frequency Spectrum (dB) Anechoic Reverberant Right Ear 50 100 150 200 0.1 0.5 1 5 10 Time (ms) Frequency (kHz)

Magnitude Spectrum (dB) Reverberation Effects Depend on Source Direction, Increase with Distance 0˚ azimuth 90˚ azimuth Left Right Left Right 15 cm Anechoic Anechoic Reverberant Reverberant Magnitude Spectrum (dB) 200 cm 0.1 1 10 0.1 1 10 0.1 1 10 0.1 1 10 Frequency (kHz)

ILDs and Reverberation Source Distance Anechoic ILDs smaller: - with distance and with asymmetric reflections ILD variability: - 0 in anechoic - increase with distance - largest in Center

ITDs and Reverberation Source Distance Anechoic 0.15 m 0.40 m 1 m 1.0 0.8 0.6 X-Corr Peak Coeff 0.4 ITD of x-corr peak independent of position or distance Magnitude of x-corr peak decreases with distance and with number of nearby walls Secondary x-corr peak can be larger than primary 0.2 0.0 0.8 0.6 ITD (ms) 0.4 0.2 0.0 30 60 90 30 60 90 30 60 90 Source Azimuth (degs)

Reverberation and ITD Variation Anechoic baseline Variation is random around mean in Center (and Back ) In Corner (and Ear) the early asymmetric reflections are have more pronounced effect 0.8 0.6 ITD (ms) 0.4 0.2 0.0 -0.2 -0.4 -0.6 -0.8 200 600 1000 1400 1800 200 600 1000 1400 1800 Frequency (Hz)

Human performance Two significant trends: 1) In Back and Corner, bias towards median plane (3.5°) - no correlation with acoustics where Corner and Ear similar 2) Variance in azimuth increases from Center to Corner - consistent with variability in ITD and ILD - however, not consistent as a function of distance

Conclusions and discussion All localization cues affected strongly by reverberation in a room position dependent way Localization performance affected modestly Possible explanations: humans are sensitive to changes in cues over time (precedence effect) listeners are able to estimate effect of reverberation and compensate for it Future work: apply models of auditory processing (Colburn, 1977) to predict how cues are extracted from reverberant signals in the brain look at effects of learning (Kopčo et al., 2001) on localization in rooms