Presented By: Shamil. C Roll no: 68 E.I Guided By: Asif Ali Lecturer in E.I.

Slides:



Advertisements
Similar presentations
Signal Estimation Technology Inc. Maher S. Maklad Optimal Resolution of Noisy Seismic Data Resolve++
Advertisements

Advanced Speech Enhancement in Noisy Environments
Hierarchy of Design Voice Controlled Remote Voice Input Control Path Speech Processing IR Interface.
2004 COMP.DSP CONFERENCE Survey of Noise Reduction Techniques Maurice Givens.
Volkan Cevher, Marco F. Duarte, and Richard G. Baraniuk European Signal Processing Conference 2008.
Model-Based Fusion of Bone and Air Sensors for Speech Enhancement and Robust Speech Recognition John Hershey, Trausti Kristjansson, Zhengyou Zhang, Alex.
Advances in WP1 Nancy Meeting – 6-7 July
Communications & Multimedia Signal Processing Meeting 6 Esfandiar Zavarehei Department of Electronic and Computer Engineering Brunel University 6 July,
Single-Channel Speech Enhancement in Both White and Colored Noise Xin Lei Xiao Li Han Yan June 5, 2002.
Speech Enhancement Based on a Combination of Spectral Subtraction and MMSE Log-STSA Estimator in Wavelet Domain LATSI laboratory, Department of Electronic,
3/24/2006Lecture notes for Speech Communications Multi-channel speech enhancement Chunjian Li DICOM, Aalborg University.
Speech Recognition in Noise
Communications & Multimedia Signal Processing Formant Tracking LP with Harmonic Plus Noise Model of Excitation for Speech Enhancement Qin Yan Communication.
Advances in WP1 and WP2 Paris Meeting – 11 febr
Matched Filters By: Andy Wang.
Gaussian Mixture-Sound Field Landmark Model for Robot Localization Talker: Prof. Jwu-Sheng Hu Department of Electrical and Control Engineering National.
THALES RESEARCH & TECHNOLOGY FRANCE This document and any data included are the property of THALES. They cannot be reproduced, disclosed or used without.
A Full Frequency Masking Vocoder for Legal Eavesdropping Conversation Recording R. F. B. Sotero Filho, H. M. de Oliveira (qPGOM), R. Campello de Souza.
On the Accuracy of Modal Parameters Identified from Exponentially Windowed, Noise Contaminated Impulse Responses for a System with a Large Range of Decay.
Normalization of the Speech Modulation Spectra for Robust Speech Recognition Xiong Xiao, Eng Siong Chng, and Haizhou Li Wen-Yi Chu Department of Computer.
A VOICE ACTIVITY DETECTOR USING THE CHI-SQUARE TEST
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Introduction SNR Gain Patterns Beam Steering Shading Resources: Wiki:
Microphone Integration – Can Improve ARS Accuracy? Tom Houy
POWER CONTROL IN COGNITIVE RADIO SYSTEMS BASED ON SPECTRUM SENSING SIDE INFORMATION Karama Hamdi, Wei Zhang, and Khaled Ben Letaief The Hong Kong University.
Speech Enhancement Using Spectral Subtraction
REVISED CONTEXTUAL LRT FOR VOICE ACTIVITY DETECTION Javier Ram’ırez, Jos’e C. Segura and J.M. G’orriz Dept. of Signal Theory Networking and Communications.
1.Processing of reverberant speech for time delay estimation. Probleme: -> Getting the time Delay of a reverberant speech with severals microphone. ->Getting.
WDM Fiber Vibrometry Krzysztof M. Abramski.
Experimental Results ■ Observations:  Overall detection accuracy increases as the length of observation window increases.  An observation window of 100.
Nico De Clercq Pieter Gijsenbergh.  Problem  Solutions  Single-channel approach  Multichannel approach  Our assignment Overview.
Authors: Sriram Ganapathy, Samuel Thomas, and Hynek Hermansky Temporal envelope compensation for robust phoneme recognition using modulation spectrum.
Peng Lei Beijing University of Aeronautics and Astronautics IGARSS 2011, Vancouver, Canada July 26, 2011 Radar Micro-Doppler Analysis and Rotation Parameter.
Noise Reduction Two Stage Mel-Warped Weiner Filter Approach.
Speech Enhancement Using a Minimum Mean Square Error Short-Time Spectral Amplitude Estimation method.
대화형 인터페이스 제안서 팀명 : Noise Suppression 팀원 : 김세희, 이호용, 서재필.
Study of Broadband Postbeamformer Interference Canceler Antenna Array Processor using Orthogonal Interference Beamformer Lal C. Godara and Presila Israt.
Full-rank Gaussian modeling of convolutive audio mixtures applied to source separation Ngoc Q. K. Duong, Supervisor: R. Gribonval and E. Vincent METISS.
IIT Bombay ICSCI 2004, Hyderabad, India, Feb’ 04 Introduction Analysis / synthesis Spec. Sub. Methodology Results Conclusion and.
AY-REU Project 7: Functional Model of a Cadaveric Human Middle Ear William Holden, Senior Chemical Engineering Margaret Welch, Pre-Junior Biomedical Engineering.
Voice Activity Detection based on OptimallyWeighted Combination of Multiple Features Yusuke Kida and Tatsuya Kawahara School of Informatics, Kyoto University,
EEL 6586: AUTOMATIC SPEECH PROCESSING Speech Features Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida February 27,
Present document contains informations proprietary to France Telecom. Accepting this document means for its recipient he or she recognizes the confidential.
Spectrum Sensing In Cognitive Radio Networks
1 Introduction1 Introduction 2 Noise red. tech 3 Spect. Subtr. 4. QBNE 5 Invest. QBNE 6 Conc., & future work2 Noise red. tech 3 Spect. Subtr.4. QBNE5 Invest.
IIT Bombay 17 th National Conference on Communications, Jan. 2011, Bangalore, India Sp Pr. 1, P3 1/21 Detection of Burst Onset Landmarks in Speech.
A. R. Jayan, P. C. Pandey, EE Dept., IIT Bombay 1 Abstract Perception of speech under adverse listening conditions may be improved by processing it to.
Statistical Signal Processing Research Laboratory(SSPRL) UT Acoustic Laboratory(UTAL) A TWO-STAGE DATA-DRIVEN SINGLE MICROPHONE SPEECH ENHANCEMENT WITH.
Analysis of Traction System Time-Varying Signals using ESPRIT Subspace Spectrum Estimation Method Z. Leonowicz, T. Lobos
Suppression of Musical Noise Artifacts in Audio Noise Reduction by Adaptive 2D filtering Alexey Lukin AES Member Moscow State University, Moscow, Russia.
1 Introduction1 Introduction 2 Spectral subtraction 3 QBNE 4 Results 5 Conclusion, & future work2 Spectral subtraction 3 QBNE4 Results5 Conclusion, & future.
Benedikt Loesch and Bin Yang University of Stuttgart Chair of System Theory and Signal Processing International Workshop on Acoustic Echo and Noise Control,
Variable Step-Size Adaptive Filters for Acoustic Echo Cancellation Constantin Paleologu Department of Telecommunications
Statistical Signal Processing Research Laboratory(SSPRL) UT Acoustic Laboratory(UTAL) SINGLE CHANNEL SPEECH ENHANCEMENT TECHNIQUE FOR LOW SNR QUASI-PERIODIC.
语音与音频信号处理研究室 Speech and Audio Signal Processing Lab Multiplicative Update of AR gains in Codebook- driven Speech.
UNIT-IV. Introduction Speech signal is generated from a system. Generation is via excitation of system. Speech travels through various media. Nature of.
PART II: TRANSIENT SUPPRESSION. IntroductionIntroduction Cohen, Gannot and Talmon\11 2 Transient Interference Suppression Transient Interference Suppression.
Institute for Experimental Mathematics Ellernstrasse Essen - Germany DATA COMMUNICATION introduction A.J. Han Vinck May 10, 2003.
HIGH-RESOLUTION SINUSOIDAL MODELING OF UNVOICED SPEECH GEORGE P. KAFENTZIS, YANNIS STYLIANOU MULTIMEDIA INFORMATICS LABORATORY DEPARTMENT OF COMPUTER SCIENCE.
Voice Activity Detection Based on Sequential Gaussian Mixture Model Zhan Shen, Jianguo Wei, Wenhuan Lu, Jianwu Dang Tianjin Key Laboratory of Cognitive.
Speech Enhancement Summer 2009
Spectral Analysis Spectral analysis is concerned with the determination of the energy or power spectrum of a continuous-time signal It is assumed that.
Speech Enhancement with Binaural Cues Derived from a Priori Codebook
Radar Micro-Doppler Analysis and Rotation Parameter Estimation for Rigid Targets with Complicated Micro-Motions Peng Lei, Jun Wang, Jinping Sun Beijing.
朝陽科技大學 資訊工程系 謝政勳 Application of GM(1,1) Model to Speech Enhancement and Voice Activity Detection 朝陽科技大學 資訊工程系 謝政勳
Lecture 7. Learning (IV): Error correcting Learning and LMS Algorithm
Microphone Array Project
Dealing with Acoustic Noise Part 1: Spectral Estimation
Presenter: Shih-Hsiang(士翔)
Microphone array Raymond Sastraputera.
Combination of Feature and Channel Compensation (1/2)
Presentation transcript:

Presented By: Shamil. C Roll no: 68 E.I Guided By: Asif Ali Lecturer in E.I

 Introduction  Speech measurement with LDV  Principe of LDV  Measurement Setup  Problem formulation  Speech Enhancement Algorithm  Speckle noise suppression  LDV-Based time frequency VAD  Spectral gain modification  Experimental Results  Conclusion

 Achieving high speech intelligibility in noisy environments is one of the most challenging and important problems for existing speech- enhancement and speech-recognition systems.  Recently, several approaches have been proposed that make use of auxiliary non acoustic sensors, such as bone and throat- microphones.  Major drawback of most existing sensors is the requirement for a physical contact between the sensor and the speaker.  Here present an alternative approach that enables a remote measurement of speech, using an auxiliary laser Doppler vibrometer (LDV) sensor.

 fd(t) = 2ν(t) cos(α)/λ ν(t)=> instantaneous throat-vibrational velocity α => Angle between the object beam and the velocity vector λ =>laser wavelength.  LDV-output signal after an FM-demodulator is Z(t) = f b + [2Av cos(α)/λ].cos(2πf v t). (1)

 Employing the VibroMet™500V LDV.  Consists of a remote laser-sensor head and an electronic controller.  Operates at 780 nm wavelength.  Can detect vibration frequencies from DC to over 40 kHz.  Its operational working distance ranges from 1 cm to 5 m.

 let y(n) =x(n) + d(n) y(n)-observed signal in the acoustic sensor. x(n) -Speech signal. d(n)-Un correlated additive noise signal.  In the STFT domain, Y lk = X lk + D lk Where l= 0, 1,... is the frame index. k = 0, 1,..., N − 1is the frequency- bin index.

Use overlapping frames of N samples with a framing-step of M samples. Let H 0lk and H 1lk indicate, respectively, speech absence and presence hypotheses in the time-frequency bin (l, k), i.e., H 0lk : Y lk = D lk H 1lk : Y lk = X lk + D lk. X̂ lk = G lk Y lk.

 The OM-LSA estimator minimizes the log spectral amplitude under signal presence uncertainty resulting in, G lk = {G H1lk }ˆP lk.G min ˆ1 − P lk. Where, G H1lk is a conditional gain function given H 1lk & G min << 1 is a constant attenuation factor. P lk is the conditional speech presence probability.

 Denoting by ξlk and γlk we get, is the a priori probability for speech absence, -Posteriori SNR -Priori SNR

 Speckle-Noise Suppression The output of the speckle-noise detector is, W l (n) = G l Z l (n) Where G l = Gs min <<1 for I l = 1(speckle noise is present) G l = 1 otherwise.

-Represents the noise-estimate bias -Smoothed-version of the power spectrum Then, we propose the following soft- decision VAD:

Speech in a given frame is defined by We attenuate high-energy transient components to the level of the stationary background noise by updating the gain floor to -Stationary noise-spectrum estimate -Smoothed noisy spectrum

 Speckle noise was successfully attenuated from the LDV-measured signal using a kurtosis-based decision rule.  A soft-decision VAD was derived in the time-frequency domain and the gain function of the OM-LSA algorithm was appropriately modified.  The effectiveness of the proposed approach in suppressing highly non-stationary noise components was demonstrated.

 I. Cohen and B. Berdugo, “Speech enhancement for nonstationary noise environment,” Signal Process., vol. 81  T. F. Quatieri, K. Brady, D. Messing, J. P. Campbell, W. M. Campbell, M. S. Brandstein, C. J.Weinstein, J. D. Tardelli, and P. D. Gatewood, “Exploiting nonacoustic sensors for speech encoding,”  T. Dekens, W. Verhelst, F. Capman, and F. Beaugendre, “Improved speech recognition in noisy environments by using a throat microphone for accurate voicing detection,” in 18th European Signal Processing Conf. (EUSIPCO), Aallborg, Denmark, Aug. 2010, pp. 23–27  M. Johansmann, G. Siegmund, and M. Pineda, “Targeting the limits of laser doppler vibrometry,” 