AN EXPECTATION MAXIMIZATION APPROACH FOR FORMANT TRACKING USING A PARAMETER-FREE NON-LINEAR PREDICTOR Issam Bazzi, Alex Acero, and Li Deng Microsoft Research.

Slides:



Advertisements
Similar presentations
Robust Speech recognition V. Barreaud LORIA. Mismatch Between Training and Testing n mismatch influences scores n causes of mismatch u Speech Variation.
Advertisements

1 A Spectral-Temporal Method for Pitch Tracking Stephen A. Zahorian*, Princy Dikshit, Hongbing Hu* Department of Electrical and Computer Engineering Old.
Speech & Audio Coding TSBK01 Image Coding and Data Compression Lecture 11, 2003 Jörgen Ahlberg.
Speech Recognition with Hidden Markov Models Winter 2011
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
Pitch Prediction From MFCC Vectors for Speech Reconstruction Xu shao and Ben Milner School of Computing Sciences, University of East Anglia, UK Presented.
Itay Ben-Lulu & Uri Goldfeld Instructor : Dr. Yizhar Lavner Spring /9/2004.
G.S.MOZE COLLEGE OF ENGINNERING BALEWADI,PUNE -45.
Speech Group INRIA Lorraine
VOICE CONVERSION METHODS FOR VOCAL TRACT AND PITCH CONTOUR MODIFICATION Oytun Türk Levent M. Arslan R&D Dept., SESTEK Inc., and EE Eng. Dept., Boğaziçi.
Speech and Audio Processing and Recognition
6/3/20151 Voice Transformation : Speech Morphing Gidon Porat and Yizhar Lavner SIPL – Technion IIT December
Model-Based Fusion of Bone and Air Sensors for Speech Enhancement and Robust Speech Recognition John Hershey, Trausti Kristjansson, Zhengyou Zhang, Alex.
Page 0 of 8 Time Series Classification – phoneme recognition in reconstructed phase space Sanjay Patil Intelligent Electronics Systems Human and Systems.
Communications & Multimedia Signal Processing Meeting 7 Esfandiar Zavarehei Department of Electronic and Computer Engineering Brunel University 23 November,
Speaker Adaptation for Vowel Classification
EE2F1 Speech & Audio Technology Sept. 26, 2002 SLIDE 1 THE UNIVERSITY OF BIRMINGHAM ELECTRONIC, ELECTRICAL & COMPUTER ENGINEERING Digital Systems & Vision.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Feature vs. Model Based Vocal Tract Length Normalization for a Speech Recognition-based Interactive Toy Jacky CHAU Department of Computer Science and Engineering.
Communications & Multimedia Signal Processing Refinement in FTLP-HNM system for Speech Enhancement Qin Yan Communication & Multimedia Signal Processing.
Voice Transformation Project by: Asaf Rubin Michael Katz Under the guidance of: Dr. Izhar Levner.
HIWIRE Progress Report – July 2006 Technical University of Crete Speech Processing and Dialog Systems Group Presenter: Alex Potamianos Technical University.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Isolated-Word Speech Recognition Using Hidden Markov Models
„Bandwidth Extension of Speech Signals“ 2nd Workshop on Wideband Speech Quality in Terminals and Networks: Assessment and Prediction 22nd and 23rd June.
1 CSE 552/652 Hidden Markov Models for Speech Recognition Spring, 2006 Oregon Health & Science University OGI School of Science & Engineering John-Paul.
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
Speech Coding Using LPC. What is Speech Coding  Speech coding is the procedure of transforming speech signal into more compact form for Transmission.
Page 0 of 23 MELP Vocoders Nima Moghadam SN#: Saeed Nari SN#: Supervisor Dr. Saameti April 2005 Sharif University of Technology.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Speech Coding Submitted To: Dr. Mohab Mangoud Submitted By: Nidal Ismail.
1 Linear Prediction. 2 Linear Prediction (Introduction) : The object of linear prediction is to estimate the output sequence from a linear combination.
1 Linear Prediction. Outline Windowing LPC Introduction to Vocoders Excitation modeling  Pitch Detection.
♥♥♥♥ 1. Intro. 2. VTS Var.. 3. Method 4. Results 5. Concl. ♠♠ ◄◄ ►► 1/181. Intro.2. VTS Var..3. Method4. Results5. Concl ♠♠◄◄►► IIT Bombay NCC 2011 : 17.
Speech Signal Representations I Seminar Speech Recognition 2002 F.R. Verhage.
Speaker Recognition by Habib ur Rehman Abdul Basit CENTER FOR ADVANCED STUDIES IN ENGINERING Digital Signal Processing ( Term Project )
Recognition of Speech Using Representation in High-Dimensional Spaces University of Washington, Seattle, WA AT&T Labs (Retd), Florham Park, NJ Bishnu Atal.
Linear Predictive Analysis 主講人:虞台文. Contents Introduction Basic Principles of Linear Predictive Analysis The Autocorrelation Method The Covariance Method.
ECE 5525 Osama Saraireh Fall 2005 Dr. Veton Kepuska
Singer similarity / identification Francois Thibault MUMT 614B McGill University.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
Speaker Identification by Combining MFCC and Phase Information Longbiao Wang (Nagaoka University of Technologyh, Japan) Seiichi Nakagawa (Toyohashi University.
Performance Comparison of Speaker and Emotion Recognition
ICASSP 2007 Robustness Techniques Survey Presenter: Shih-Hsiang Lin.
More On Linear Predictive Analysis
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Cameron Rowe.  Introduction  Purpose  Implementation  Simple Example Problem  Extended Kalman Filters  Conclusion  Real World Examples.
By Sarita Jondhale 1 Signal preprocessor: “conditions” the speech signal s(n) to new form which is more suitable for the analysis Postprocessor: operate.
Linear Prediction.
Feature Transformation and Normalization Present by Howard Reference : Springer Handbook of Speech Processing, 3.3 Environment Robustness (J. Droppo, A.
1 Speech Compression (after first coding) By Allam Mousa Department of Telecommunication Engineering An Najah University SP_3_Compression.
CPH Dr. Charnigo Chap. 11 Notes Figure 11.2 provides a diagram which shows, at a glance, what a neural network does. Inputs X 1, X 2,.., X P are.
Yun, Hyuk Jin. Theory A.Nonuniformity Model where at location x, v is the measured signal, u is the true signal emitted by the tissue, is an unknown.
Biointelligence Laboratory, Seoul National University
High Quality Voice Morphing
LECTURE 11: Advanced Discriminant Analysis
ARTIFICIAL NEURAL NETWORKS
Vocoders.
Ch3: Model Building through Regression
Department of Civil and Environmental Engineering
Linear Prediction.
1 Vocoders. 2 The Channel Vocoder (analyzer) : The channel vocoder employs a bank of bandpass filters,  Each having a bandwidth between 100 HZ and 300.
Voice conversion using Artificial Neural Networks
The Vocoder and its related technology
Vocoders.
Foundation of Video Coding Part II: Scalar and Vector Quantization
EE513 Audio Signals and Systems
Linear Prediction.
Combination of Feature and Channel Compensation (1/2)
Presentation transcript:

AN EXPECTATION MAXIMIZATION APPROACH FOR FORMANT TRACKING USING A PARAMETER-FREE NON-LINEAR PREDICTOR Issam Bazzi, Alex Acero, and Li Deng Microsoft Research One Microsoft Way Redmond, WA, USA 2003

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20072/14 Outline Introduction The Model EM Training Format Tracking Experiment Results Conclusion

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20073/14 Introduction Traditional methods use LPC or matching stored templates of spectral cross sections In either case, formant tracking is error-prone due to not enough candidates or templates This paper uses a predictor codebook of MFCC to present formant relationships Also, this method explores the complete formant space, avoiding premature elimination in LPC or template matching

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20074/14 The Model o t = F(x t ) + r t o t is observed MFCC coefficients x t is vocal tract resonances (VTR) and corresponding bandwidths F(x t ) is the quantized frequency and bandwidth of formants, named predictor codebook r t is the residual signal

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20075/14 Constructing F(x) All-pole model Assume there are I formants x = (F 1, B 1, F 2, B 2,……, F I, B I ) Then use z-transfrom to get H(z): Finally, each quantized VTR x can be transformed into a MFCC series F(x)

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20076/14 EM Training (1/2) Use a single Gaussian to model r t T frames utterance, θ is parameters (mean and covariance) of Gaussian Assume formant values x are uniformly distributed, and can take any of C quantized values

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20077/14 EM Training (2/2)

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20078/14 Formant Tracking (1/2) Frame-by-Frame Tracking  Formants in each frame are estimated independently  One-to-one Mapping (MAP)  Minimum Mean Squared Error (MMSE)

Ts'ai, Chung-Ming, Speech Lab, NTUST, 20079/14 Formant Tracking (2/2) Tracking with Continuity Constraints  First Order State Model: x t = x t-1 + w t  w t is modeled as a Gaussian with zero mean and diagonal Σ w  MAP method below can be estimated using Viterbi search  MMSE is more much complex and this paper uses an approximate method to obtain, which is not well described here

Ts'ai, Chung-Ming, Speech Lab, NTUST, /14 Experiment Settings Track 3 formants  Frequencies are first mapped on mel-scale then uniformly quantized  Bandwidths are simply uniformly quantized  F1 < F2 < F3, so totally entries in codebook Gain = 1 MFCC is 12 dimension, without C 0 20 utterances of one male speaker are used for EM

Ts'ai, Chung-Ming, Speech Lab, NTUST, /14 Experiment Results, “they were what”

Ts'ai, Chung-Ming, Speech Lab, NTUST, /14 Experiment Results, with bandwidth

Ts'ai, Chung-Ming, Speech Lab, NTUST, /14 Experiment Results, residual

Ts'ai, Chung-Ming, Speech Lab, NTUST, /14 Conclusion This method is totally unsupervised, needless of any labeling Works well in unvoiced frames No gross errors May be applied to speech recognizing system