1 Brain Neuroinformatics: Auditory System February 27, 2002 Soo-Young Lee Brain Science Research Center Korea Advanced Institute of Science & Technology.

Slides:



Advertisements
Similar presentations
Acoustic Echo Cancellation for Low Cost Applications
Advertisements

Figures for Chapter 7 Advanced signal processing Dillon (2001) Hearing Aids.
Robust Speech recognition V. Barreaud LORIA. Mismatch Between Training and Testing n mismatch influences scores n causes of mismatch u Speech Variation.
Advances in WP1 Trento Meeting January
Speech Enhancement through Noise Reduction By Yating & Kundan.
Combining Heterogeneous Sensors with Standard Microphones for Noise Robust Recognition Horacio Franco 1, Martin Graciarena 12 Kemal Sonmez 1, Harry Bratt.
Advanced Speech Enhancement in Noisy Environments
Multipitch Tracking for Noisy Speech
PCS Research & Advanced Technology Labs Speech Lab How to deal with the noise in real systems? Hsiao-Chun Wu Motorola PCS Research and Advanced Technology.
2004 COMP.DSP CONFERENCE Survey of Noise Reduction Techniques Maurice Givens.
Advances in WP1 Turin Meeting – 9-10 March
Signal Processing Institute Swiss Federal Institute of Technology, Lausanne 1 Feature selection for audio-visual speech recognition Mihai Gurban.
Blind Source Separation of Acoustic Signals Based on Multistage Independent Component Analysis Hiroshi SARUWATARI, Tsuyoki NISHIKAWA, and Kiyohiro SHIKANO.
3/24/2006Lecture notes for Speech Communications Multi-channel speech enhancement Chunjian Li DICOM, Aalborg University.
Subband-based Independent Component Analysis Y. Qi, P.S. Krishnaprasad, and S.A. Shamma ECE Department University of Maryland, College Park.
HIWIRE MEETING CRETE, SEPTEMBER 23-24, 2004 JOSÉ C. SEGURA LUNA GSTC UGR.
Speech Recognition in Noise
Communications & Multimedia Signal Processing Refinement in FTLP-HNM system for Speech Enhancement Qin Yan Communication & Multimedia Signal Processing.
Advances in WP1 and WP2 Paris Meeting – 11 febr
1 New Technique for Improving Speech Intelligibility for the Hearing Impaired Miriam Furst-Yust School of Electrical Engineering Tel Aviv University.
Why is ASR Hard? Natural speech is continuous
Analytical Techniques
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Adaptive Noise Cancellation ANC W/O External Reference Adaptive Line Enhancement.
Speech Segregation Based on Sound Localization DeLiang Wang & Nicoleta Roman The Ohio State University, U.S.A. Guy J. Brown University of Sheffield, U.K.
Normalization of the Speech Modulation Spectra for Robust Speech Recognition Xiong Xiao, Eng Siong Chng, and Haizhou Li Wen-Yi Chu Department of Computer.
A VOICE ACTIVITY DETECTOR USING THE CHI-SQUARE TEST
1 7-Speech Recognition (Cont’d) HMM Calculating Approaches Neural Components Three Basic HMM Problems Viterbi Algorithm State Duration Modeling Training.
Microphone Integration – Can Improve ARS Accuracy? Tom Houy
Scheme for Improved Residual Echo Cancellation in Packetized Audio Transmission Jivesh Govil Digital Signal Processing Laboratory Department of Electronics.
7-Speech Recognition Speech Recognition Concepts
Nico De Clercq Pieter Gijsenbergh Noise reduction in hearing aids: Generalised Sidelobe Canceller.
Page 0 of 23 MELP Vocoders Nima Moghadam SN#: Saeed Nari SN#: Supervisor Dr. Saameti April 2005 Sharif University of Technology.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Reporter: Shih-Hsiang( 士翔 ). Introduction Speech signal carries information from many sources –Not all information is relevant or important for speech.
Compensating speaker-to-microphone playback system for robust speech recognition So-Young Jeong and Soo-Young Lee Brain Science Research Center and Department.
LOG-ENERGY DYNAMIC RANGE NORMALIZATON FOR ROBUST SPEECH RECOGNITION Weizhong Zhu and Douglas O’Shaughnessy INRS-EMT, University of Quebec Montreal, Quebec,
Implementation of Infomax ICA Algorithm with Low-Power Analog CMOS Circuits Ki-Seok Cho and Soo-Young Lee Brain Science Research Center and Department.
Basics of Neural Networks Neural Network Topologies.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
Yi-zhang Cai, Jeih-weih Hung 2012/08/17 報告者:汪逸婷 1.
1 Robust Endpoint Detection and Energy Normalization for Real-Time Speech and Speaker Recognition Qi Li, Senior Member, IEEE, Jinsong Zheng, Augustine.
Gammachirp Auditory Filter
A NEW FEATURE EXTRACTION MOTIVATED BY HUMAN EAR Amin Fazel Sharif University of Technology Hossein Sameti, S. K. Ghiathi February 2005.
Independent Component Analysis Algorithm for Adaptive Noise Cancelling 적응 잡음 제거를 위한 독립 성분 분석 알고리즘 Hyung-Min Park, Sang-Hoon Oh, and Soo-Young Lee Brain.
Robust Entropy-based Endpoint Detection for Speech Recognition in Noisy Environments 張智星
Hearing Research Center
A Model of Binaural Processing Based on Tree-Structure Filter-Bank
Laboratory for Experimental ORL K.U.Leuven, Belgium Dept. of Electrotechn. Eng. ESAT/SISTA K.U.Leuven, Belgium Combining noise reduction and binaural cue.
Speech Enhancement for ASR by Hans Hwang 8/23/2000 Reference 1. Alan V. Oppenheim,etc., ” Multi-Channel Signal Separation by Decorrelation ”,IEEE Trans.
Hearing Sound and the limits to hearing Structure of the ear: Outer, middle, inner Outer ear and middle ear functions Inner ear: the cochlea - Frequency.
Microphone Array Project ECE5525 – Speech Processing Robert Villmow 12/11/03.
Performance Comparison of Speaker and Emotion Recognition
RCC-Mean Subtraction Robust Feature and Compare Various Feature based Methods for Robust Speech Recognition in presence of Telephone Noise Amin Fazel Sharif.
Independent Component Analysis Independent Component Analysis.
Chapter 7 Speech Recognition Framework  7.1 The main form and application of speech recognition  7.2 The main factors of speech recognition  7.3 The.
SOME SIMPLE MANIPULATIONS OF SOUND USING DIGITAL SIGNAL PROCESSING Richard M. Stern demo January 15, 2015 Department of Electrical and Computer.
Spatial vs. Blind Approaches for Speaker Separation: Structural Differences and Beyond Julien Bourgeois RIC/AD.
January 2001RESPITE workshop - Martigny Multiband With Contaminated Training Data Results on AURORA 2 TCTS Faculté Polytechnique de Mons Belgium.
Motorola presents in collaboration with CNEL Introduction  Motivation: The limitation of traditional narrowband transmission channel  Advantage: Phone.
1 7-Speech Recognition Speech Recognition Concepts Speech Recognition Approaches Recognition Theories Bayse Rule Simple Language Model P(A|W) Network Types.
UNIT-IV. Introduction Speech signal is generated from a system. Generation is via excitation of system. Speech travels through various media. Nature of.
A Neurodynamical Cortical Model of Visual Attention and Invariant Object Recognition Gustavo Deco Edmund T. Rolls Vision Research, 2004.
1 LOW-RESOURCE NOISE-ROBUST FEATURE POST-PROCESSING ON AURORA 2.0 Chia-Ping Chen, Jeff Bilmes and Katrin Kirchhoff SSLI Lab Department of Electrical Engineering.
Speech Enhancement Summer 2009
Speech Enhancement with Binaural Cues Derived from a Priori Codebook
The Sound of the Original Sentences
AUDIO SURVEILLANCE SYSTEMS: SUSPICIOUS SOUND RECOGNITION
INTRODUCTION TO ADVANCED DIGITAL SIGNAL PROCESSING
Presenter: Shih-Hsiang(士翔)
Combination of Feature and Channel Compensation (1/2)
Presentation transcript:

1 Brain Neuroinformatics: Auditory System February 27, 2002 Soo-Young Lee Brain Science Research Center Korea Advanced Institute of Science & Technology bsrc.kaist.ac.kr

KAIST Brain Science Research Center 2 Artificial Auditory System Based on Human Cognitive Mechanism  Develop mathematical model and auditory chip  Develop Continuous speech recognition system  Auditory model  Binaural processing  Selective attention  Continuous speech recognition 95% recognition in 10dB SNR

KAIST Brain Science Research Center 3 Auditory Pathway Superior Olive Cochlea Superior Olive Cochlea Higher-Order Brain Function (Recognition) Cochlear Nucleus Cochlear Nucleus Auditory Cortex Auditory Cortex   Improvement of Cochlea model  Binaural hearing model  Auditory cortex model  Selective attention model  Continuous-speech recognition system  Speech-recognition chip

KAIST Brain Science Research Center 4 Object Path Attention Spatial Path Higher- Order Functions Top-Down Attention Bottom-Up Attention TD/BU Attention Cross Correlation Sound Localization Speech Enhancement Speaker Separation Nonlinear Features Time Adaptation Complex Sound Time-Freq. Masking Speech Rec. System Parser Cochlea Aud. Cortex Research Scopes Auditory Pathway

KAIST Brain Science Research Center 5 Masking - Lateral Inhibition - Recursive

KAIST Brain Science Research Center 6 Masking using Lateral Inhibition

KAIST Brain Science Research Center 7 Temporal Masking

KAIST Brain Science Research Center 8 Temporal Masking: Integration Model

KAIST Brain Science Research Center 9 Time & Frequency Response

KAIST Brain Science Research Center 10 Isolated Word Recognition Rates

KAIST Brain Science Research Center 11 Discussions: Masking Masking Suppresses Unwanted Noisy Components in Signal Simultaneous Masking by Lateral Inhibition - Recognition performance was enhanced with MFCC model - Proposed model can be used with any auditory model - ZCPA has spectral masking effect Temporal Masking by Unilateral Inhibition - Unilateral inhibition using the integration model - Model resembles other feature processing algorithms - Recognition performance was enhanced with RASTA parameters

KAIST Brain Science Research Center 12 ASR in mismatched environments Environmental information Background noise, acoustic/transmission channel Assume environment degradation model Channel Compensation Clean speech Channel Additive noise Distorted speech

KAIST Brain Science Research Center 13 Mapper train Where and which type of mapper should be deployed? Speaker-to-Microphone mapping F.E. + clean F.E.Mapper distorted Error F.E.Trained Mapper distorted To recognizer

KAIST Brain Science Research Center 14 Adaptive Noise Cancelling Adaptive noise cancelling An approach to reduce noise based on reference noise signals System output The LMS algorithm

KAIST Brain Science Research Center 15 ICA-based Approach to ANC The difference between the LMS algorithm and the ICA-based approach Existence of the score function The LMS algorithm Decorrelate output signal from the reference input The ICA-based approach Make output signal independent of the reference input Independence Involve higher-order statistics including correlation The ICA-based approach Remove the noise components using higher-order statistics and correlation

KAIST Brain Science Research Center 16 TDAF approach to ANC Normalized LMS algorithm Normalized ICA-based algorithm where

KAIST Brain Science Research Center 17 Experimental Results (1) Experiments for artificially generated i.i.d. signals SNRs of output signals for the simple simulation mixing filter (dB) Signal and Noise Initial SNRs SNRs after convergence LMS algorithm ICA-based approach Laplacian Gaussian

KAIST Brain Science Research Center 18 Experimental Results (2) Experiments for recorded signals Signal waveforms for the car noise and the simple simulation filter Signal sourceNoise source Primary input signalSystem output signal

KAIST Brain Science Research Center 19 Experimental Results (3) Experiments for recorded signals SNRs of output signals for the measured filter

KAIST Brain Science Research Center 20 Experimental Results (4) Comparison of learning curves with and without TDAF Car noise The ICA-based approachThe LMS algorithm

KAIST Brain Science Research Center 21 Discussion: ANC A method to ANC based on ICA was proposed. The ICA-based learning rule was derived. The ICA-based approach Include higher-order statistics Make the output independent of the reference input The LMS algorithm Make the output uncorrelated to the reference input Gave better performances than the LMS algorithm TDAF method was applied to the ICA-based approach. Derived the normalized ICA-based learning rule Improved convergence rates

KAIST Brain Science Research Center 22 Bottom-Up and Top-Down Attention Attended Output Top-Down Expectation Bottom-Up Recognition Environment External Cue Brain Classifier Output Internal Cue Input Features Attended Input Input Stimulus Bottom-Up Attention  Bottom-Up: - Masking - ICA  Top-Down: - MLP - HMM

KAIST Brain Science Research Center 23 디지털 음성 인식 프로세서 (1) 실시간 음성 인식 프로세서 인텔 8051 프로세서 채택 ( 범용성 ) 12 MHz 동작 : 1.93 MIPS 하드웨어 특징 추출 (16 채널 ) 고속 인식을 위한 인식 가속기 50 단어 시 20 ms 이내 인식 결과 출력 인식률 95% 이상

KAIST Brain Science Research Center 24 디지털 음성 인식 프로세서 (2) Hynix 0.35 공정 패키지 형태 : 64-TQFP 내부 메모리 : 128x12, 256x8, 2048x8 AGC, A/D, D/A 내장 12KHz, 12 bit A/D 12Khz, 8 bit D/A 인식을 위한 특수 레지스터 특징 추출 레지스터 인식 가속 레지스터 인식을 위한 효율적인 구조 AGCA/DD/A 2048x8 128X12 256X8 CM8051

KAIST Brain Science Research Center 25 디지털 음성 인식 프로세서 (3) CM8051 블록도 및 I/O 구성도

KAIST Brain Science Research Center 26 디지털 음성 인식 프로세서 (4) 기본 구성 회로도

KAIST Brain Science Research Center 27 Analogue ICA Chip   multiplier   current summation r  learning rate 4 x 4 ICA network in one Chip

KAIST Brain Science Research Center 28 Test Results in Waveforms Source Signals (s1, s2) two different male’s voice 16 kHz sampled Mixed Signals (x1, x2) Instantaneous mixture Mixing Mtx A is Separated Signals (o1, o2) Recovered original sources

KAIST Brain Science Research Center 29 Fabricated Chip 2.8mm x 2.8mm AMS CMOS 0.6um 2 poly-3 metal Analog Digital Hybrid Process Die Photo of a Fabricated ICA Chip