LINEAR DYNAMIC MODEL FOR CONTINUOUS SPEECH RECOGNITION URL: Ph.D.

Slides:



Advertisements
Similar presentations
Robust Speech recognition V. Barreaud LORIA. Mismatch Between Training and Testing n mismatch influences scores n causes of mismatch u Speech Variation.
Advertisements

Building an ASR using HTK CS4706
Speech Recognition Part 3 Back end processing. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
Acoustic Model Adaptation Based On Pronunciation Variability Analysis For Non-Native Speech Recognition Yoo Rhee Oh, Jae Sam Yoon, and Hong Kook Kim Dept.
Dual-domain Hierarchical Classification of Phonetic Time Series Hossein Hamooni, Abdullah Mueen University of New Mexico Department of Computer Science.
PERFORMANCE ANALYSIS OF AURORA LARGE VOCABULARY BASELINE SYSTEM Naveen Parihar, and Joseph Picone Center for Advanced Vehicular Systems Mississippi State.
Motivation Traditional approach to speech and speaker recognition:
AN INVESTIGATION OF DEEP NEURAL NETWORKS FOR NOISE ROBUST SPEECH RECOGNITION Michael L. Seltzer, Dong Yu Yongqiang Wang ICASSP 2013 Presenter : 張庭豪.
Hidden Markov Models in NLP
Page 0 of 8 Time Series Classification – phoneme recognition in reconstructed phase space Sanjay Patil Intelligent Electronics Systems Human and Systems.
4/25/2001ECE566 Philip Felber1 Speech Recognition A report of an Isolated Word experiment. By Philip Felber Illinois Institute of Technology April 25,
Speaker Adaptation for Vowel Classification
HIWIRE Progress Report Trento, January 2007 Presenter: Prof. Alex Potamianos Technical University of Crete Presenter: Prof. Alex Potamianos Technical University.
HIWIRE Progress Report – July 2006 Technical University of Crete Speech Processing and Dialog Systems Group Presenter: Alex Potamianos Technical University.
Why is ASR Hard? Natural speech is continuous
Nonlinear Mixture Autoregressive Hidden Markov Models for Speech Recognition S. Srinivasan, T. Ma, D. May, G. Lazarou and J. Picone Department of Electrical.
Normalization of the Speech Modulation Spectra for Robust Speech Recognition Xiong Xiao, Eng Siong Chng, and Haizhou Li Wen-Yi Chu Department of Computer.
Soft Margin Estimation for Speech Recognition Main Reference: Jinyu Li, " SOFT MARGIN ESTIMATION FOR AUTOMATIC SPEECH RECOGNITION," PhD thesis, Georgia.
HMM-BASED PSEUDO-CLEAN SPEECH SYNTHESIS FOR SPLICE ALGORITHM Jun Du, Yu Hu, Li-Rong Dai, Ren-Hua Wang Wen-Yi Chu Department of Computer Science & Information.
Statistical automatic identification of microchiroptera from echolocation calls Lessons learned from human automatic speech recognition Mark D. Skowronski.
Page 0 of 14 Dynamical Invariants of an Attractor and potential applications for speech data Saurabh Prasad Intelligent Electronic Systems Human and Systems.
1 CSE 552/652 Hidden Markov Models for Speech Recognition Spring, 2006 Oregon Health & Science University OGI School of Science & Engineering John-Paul.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
A brief overview of Speech Recognition and Spoken Language Processing Advanced NLP Guest Lecture August 31 Andrew Rosenberg.
Minimum Mean Squared Error Time Series Classification Using an Echo State Network Prediction Model Mark Skowronski and John Harris Computational Neuro-Engineering.
A Sparse Modeling Approach to Speech Recognition Based on Relevance Vector Machines Jon Hamaker and Joseph Picone Institute for.
NONLINEAR DYNAMIC INVARIANTS FOR CONTINUOUS SPEECH RECOGNITION Author: Daniel May Mississippi State University Contact Information: 1255 Louisville St.
LINEAR DYNAMIC MODEL FOR CONTINUOUS SPEECH RECOGNITION Ph.D. Candidate: Tao Ma Department of Electrical and Computer Engineering Mississippi State University.
LINEAR DYNAMIC MODEL FOR CONTINUOUS SPEECH RECOGNITION Ph.D. Candidate: Tao Ma Advised by: Dr. Joseph Picone Institute for Signal and Information Processing.
Recognition of spoken and spelled proper names Reporter : CHEN, TZAN HWEI Author :Michael Meyer, Hermann Hild.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Chun Kai Chen Author: Aravind.
LML Speech Recognition Speech Recognition Introduction I E.M. Bakker.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Experimentation Duration is the most significant feature with around 40% correlation. Experimentation Duration is the most significant feature with around.
Automatic Speech Recognition: Conditional Random Fields for ASR Jeremy Morris Eric Fosler-Lussier Ray Slyh 9/19/2008.
1 Robust Endpoint Detection and Energy Normalization for Real-Time Speech and Speaker Recognition Qi Li, Senior Member, IEEE, Jinsong Zheng, Augustine.
Speaker Verification Speaker verification uses voice as a biometric to determine the authenticity of a user. Speaker verification systems consist of two.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
PhD Candidate: Tao Ma Advised by: Dr. Joseph Picone Institute for Signal and Information Processing (ISIP) Mississippi State University Linear Dynamic.
Voice Activity Detection based on OptimallyWeighted Combination of Multiple Features Yusuke Kida and Tatsuya Kawahara School of Informatics, Kyoto University,
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
MINIMUM WORD CLASSIFICATION ERROR TRAINING OF HMMS FOR AUTOMATIC SPEECH RECOGNITION Yueng-Tien, Lo Speech Lab, CSIE National.
Presented by: Fang-Hui Chu Discriminative Models for Speech Recognition M.J.F. Gales Cambridge University Engineering Department 2007.
Automatic Speech Recognition A summary of contributions from multiple disciplines Mark D. Skowronski Computational Neuro-Engineering Lab Electrical and.
ICASSP 2007 Robustness Techniques Survey Presenter: Shih-Hsiang Lin.
S.Patil, S. Srinivasan, S. Prasad, R. Irwin, G. Lazarou and J. Picone Intelligent Electronic Systems Center for Advanced Vehicular Systems Mississippi.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
Performance Analysis of Advanced Front Ends on the Aurora Large Vocabulary Evaluation Authors: Naveen Parihar and Joseph Picone Inst. for Signal and Info.
A New Approach to Utterance Verification Based on Neighborhood Information in Model Space Author :Hui Jiang, Chin-Hui Lee Reporter : 陳燦輝.
Statistical automatic identification of microchiroptera from echolocation calls Lessons learned from human automatic speech recognition Mark D. Skowronski.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture Mark D. Skowronski Computational Neuro-Engineering Lab University of Florida March 31,
Phone-Level Pronunciation Scoring and Assessment for Interactive Language Learning Speech Communication, 2000 Authors: S. M. Witt, S. J. Young Presenter:
ASSESSING SEARCH TERM STRENGTH IN SPOKEN TERM DETECTION Amir Harati and Joseph Picone Institute for Signal and Information Processing, Temple University.
1 Experiments with Detector- based Conditional Random Fields in Phonetic Recogntion Jeremy Morris 06/01/2007.
Flexible Speaker Adaptation using Maximum Likelihood Linear Regression Authors: C. J. Leggetter P. C. Woodland Presenter: 陳亮宇 Proc. ARPA Spoken Language.
Utterance verification in continuous speech recognition decoding and training Procedures Author :Eduardo Lleida, Richard C. Rose Reporter : 陳燦輝.
Christoph Prinz / Automatic Speech Recognition Research Progress Hits the Road.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
A NONPARAMETRIC BAYESIAN APPROACH FOR
EEL 6586: AUTOMATIC SPEECH PROCESSING Hidden Markov Model Lecture
Statistical Models for Automatic Speech Recognition
HUMAN LANGUAGE TECHNOLOGY: From Bits to Blogs
EEG Recognition Using The Kaldi Speech Recognition Toolkit
Statistical Models for Automatic Speech Recognition
Automatic Speech Recognition: Conditional Random Fields for ASR
Combination of Feature and Channel Compensation (1/2)
Presentation transcript:

LINEAR DYNAMIC MODEL FOR CONTINUOUS SPEECH RECOGNITION URL: Ph.D. Proposal: Tao Ma Advised by: Dr. Joseph Picone Institute for Signal and Information Processing (ISIP) Mississippi State University January 23, 2010

Slide 1 Abstract In this dissertation work, we propose a hybrid speech recognizer to effectively integrates linear dynamic model into traditional HMM-based framework for continuous speech recognition. Traditional methods simplify speech signal as a piecewise stationary signal and speech features are assumed to be temporally uncorrelated. While these simplifications have enabled tremendous advances in speech processing systems, for the past several years progress on the core statistical models has stagnated. Recent theoretical and experimental studies suggest that exploiting frame-to-frame correlations in a speech signal further improves the performance of ASR systems. Linear Dynamic Models (LDMs) take advantage of higher order statistics or trajectories using a state space-like formulation. This smoothed trajectory model allows the system to better track the speech dynamics in noisy environments. The proposed hybrid system is capable of handling large recognition tasks such as Aurora-4 large vocabulary corpus, is robust to noise- corrupted speech data and mitigates the effort of mismatched training and evaluation conditions. This two-pass system leverages the temporal modeling and N-best list generation capabilities of the traditional HMM architecture in a first pass analysis. In the second pass, candidate sentence hypotheses are re- ranked using a phone-based LDM model.

Slide 2 Hidden Markov Models with Gaussian Mixture Models (GMMs) to model state output distributions Bayesian model based approach for speech recognition system Speech Recognition System

Slide 3 Is HMM a perfect model for speech recognition? Progress on improving the accuracy of HMM-based system has slowed in the past decade Theory drawbacks of HMM –False assumption that frames are independent and stationary –Spatial correlation is ignored (diagonal covariance matrix) –Limited discrete state space Accuracy Time Clean Noisy

Slide 4 Motivation of Linear Dynamic Model (LDM) Research Motivation –A model which reflects the characteristics of speech signals will ultimately lead to great ASR performance improvement –LDM incorporates frame correlation information of speech signals, which is potential to increase recognition accuracy –“Filter” characteristic of LDM has potential to improve noise robustness of speech recognition –Fast growing computation capacity make it realistic to build a two- way HMM/LDM hybrid speech recognizer

Slide 5 State Space Model Linear Dynamic Model (LDM) is derived from State Space Model Equations of State Space Model:

Slide 6 Equations of Linear Dynamic Model (LDM) –Current state is only determined by previous state –H, F are linear transform matrices –Epsilon and Eta are Gaussian noise components y: observation feature vector x: corresponding internal state vector H: linear transform matrix between y and x F: linear transform matrix between current state and previous state epsilon: Gaussian noise component eta: Gaussian noise component Linear Dynamic Model

Slide 7 Human Being Sound System Kalman Filtering Estimation e For a speech sound, Kalman filtering for state inference

Slide 8 Rauch-Tung-Striebel (RTS) smoother –Additional backward pass to minimize inference error –During EM training, computes the expectations of state statistics Standard Kalman FilterKalman Filter with RTS smoother RTS smoother for better inference

Slide 9 Maximum Likelihood Parameter Estimation LDM Parameters:

Slide 10 LDM for Speech Classification MFCC Feature ……… aa ch eh x y HMM-Based Recognition LDM-Based Recognition MFCC Feature ……… aa ch eh x y Hypothesis x ^ x ^ x ^ x ^ x ^ x ^ one vs. all classifier:

Slide 11 Segment-based model –frame-to-phoneme information is needed before classification EM training is sensitive to state initialization –Each phoneme is modeled by a LDM, EM training is to find a set of parameters for a specific LDM –No good mechanism for state initialization yet More parameters than HMM (2~3x) –Currently mono-phone model, to build a tri-phone model for LVCSR would need more training data Challenges of Applying LDM to ASR

Slide 12 Phoneme classification on TIDigits corpus TIDigits Corpus: more than 25 thousand digit utterances spoken by over 326 men, women, and children. dialect balanced for 21 dialectical regions of the continental U.S. Frame-to-phone alignment is generated by ISIP decoder (force align mode) 18 phones, one vs. all classifier

Slide 13 Pronunciation lexicon and broad phonetic classes WordPronunciation ZEROz iy r ow OHow ONEw ah n TWOt uw THREEth r iy FOURf ow r FIVEf ay v SIXs ih k s SEVENs eh v ih n EIGHTey t NINEn ay n PhonemeClassPhonemeClass ahVowelssFricatives ayVowelsfFricatives ehVowelsthFricatives eyVowelsvFricatives ihVowelszFricatives iyVowelswGlides uwVowelsrGlides owVowelskStops nNasalstStops Table 1: Pronunciation lexicon Table 2: Broad phonetic classes

Slide 14 Classification results for TIDigits dataset (13mfcc) The solid blue line shows classification accuracies for full covariance LDMs with state dimensions from 1 to 25. The dashed red line shows classification accuracies for diagonal covariance LDMs with state dimensions from 1 to 25. HMM baseline: 91.3% Acc; Full LDM: 91.69% Acc; Diagonal LDM: 91.66% Acc.

Slide 15 Model choice: full LDM vs. diagonal LDM Diagonal covariance LDM performs as good as full covariance LDM, with much less model parameters. Confusion phoneme pairs for the classification results using full LDMs Confusion phoneme pairs for the classification results of using diagonal LDMs

Slide 16 Classification accuracies by broad phonetic classes Classification results for fricatives and stops are high. Classification results for glides are lower (~85%). Vowels and nasals result in mediocre accuracy (89% and 93% respectively). Overall, LDMs provide a reasonably good classification performance for TIDigits.

Slide 17 Proposed work: hybrid HMM/LDM speech recognizer Motivations: LDM phoneme classification experiments provide motivation to apply it for large vocabulary, continuous speech recognition (LVCSR) system. However, developing LDM-based LVCSR system from scratch has been proved to be extremely difficult because LDM is inherently a static classifier. LDM and HMM can be complementary to each other, incorporating LDM into traditional HMM-based framework could lead to a superior system with better performance.

Slide 18 Two-pass hybrid HMM/LDM speech recognizer N-best list rescoring architecture of the hybrid recognizer Hybrid recognizer takes advantage of a HMM architecture to model the temporal evolution of speech and LDM advantages to model frame-to- frame correlation and higher order statistics. First pass: HMM generates multiple recognition hypotheses with frame-to- phoneme alignments. Second pass: incorporating LDM to re- rank the N-best sentence hypotheses and output the most possible hypothesis as the recognition result.

Slide 19 Aurora-4 corpus to evaluate hybrid recognizer Aurora-4 large vocabulary corpus is a well-established LVCSR benchmark with different noisy conditions. Acoustic Training: Derived from 5000 word WSJ0 task 16 kHz sample rate Recorded with Sennheiser microphone 83 speakers 7138 training utterances totaling in 14 hours of speech Development Sets: Derived from WSJ0 Evaluation and Development sets 7 individual test sets recorded with Sennheiser microphone Clean set plus 6 sets with noise conditions Randomly chosen SNR between 5 and 15 dB for noisy sets

Slide 20 What will be in my dissertation? Chapter 1: Introduction Chapter 2: THE STATISTICAL APPROACH FOR SPEECH RECOGNITION 2.1 The Speech Recognition Problem 2.2 Hidden Markov Models 2.3 Segment-based Models 2.4 Hybrid Connectionist Systems 2.5 Summary Chapter 3: LINEAR DYNAMIC MODELS 3.1 Linear Dynamic System 3.2 Kalman Filter 3.3 Linear Dynamic Model State Inference Model Parameter Estimation Likelihood Calculation 3.4 Summary Chapter 4: LDM FOR SPEECH CLASSIFICATION 4.1 Acoustic Front-end 4.2 TIDigits Corpus 4.3 Training from Multiple Observation Sequences 4.4 Classification Results 4.5 Summary Chapter 5: HMM/LDM ARCHITECTURE FOR SPEECH RECOGNITION 5.1 Aurora-4 Corpus 5.2 Hybrid Recognizer Architecture 5.3 Segmental Modeling 5.4 Modifications to an ASR System 5.5 N-best List Rescoring Paradigm 5.6 Experiment Results 5.7 Summary Chapter 6: CONCLUSIONS AND FUTURE DIRECTIONS REFERENCES

Slide 21 Tasks to be finished and technical risks Tasks to be finished: Validate the hybrid HMM/LDM recognizer using a small dataset to ensure correct algorithm implantation Code optimization for core LDM training and likelihood calculation Evaluate hybrid HMM/LDM speech recognizer using Aurora-4 speech corpus for both clean data and noisy data and analyze the experiment results Technical risks: LDM model training for very large dataset might lead to singular matrix problem due to arithmetic precision of matrix operation Investigation needed to optimally combine the HMM acoustic score and LDM acoustic score

Slide 22 Patents/Publications/Reports/Talks Patents P29573 Method and Apparatus for Improving Memory Locality for Real-time Speech Recognition by Michael Deisher and Tao Ma (pending patent, filed in June 2009). Publications/Reports/Talks T. Ma, S. Srinivasan, D. May, G. Lazarou and J. Picone, "Robust Speech Recognition Using Linear Dynamic Models," submitted to the IEEE Signal Processing Letters, Spring T. Ma and M. Deisher, "Novel CI-Backoff Scheme for Real-time Embedded Speech Recognition,” to be appeared in ICASSP 2010, Dallas, Texas, USA, March S. Srinivasan, T. Ma, D. May, G. Lazarou and J. Picone, "Nonlinear Statistical Modeling of Speech," presentated at the 29th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering (MaxEnt 2009), Oxford, Mississippi, USA, July S. Srinivasan, T. Ma, D. May, G. Lazarou and J. Picone, "Nonlinear Mixture Autoregressive Hidden Markov Models For Speech Recognition," Proceedings of the International Conference on Spoken Language Processing, pp , Brisbane, Australia, September T. Ma, S. Srinivasan, D. May, G. Lazarou and J. Picone, "Robust Speech Recognition Using Linear Dynamic Models,” submitted to INTERSPEECH, Brisbane, Australia, September D. May, S. Srinivasan, T. Ma and J. Picone, “Continuous Speech Recognition Using Nonlinear Dynamical Invariants,” submitted to International Conference on Acoustics, Speech, and Signal Processing, Las Vegas, Nevada, USA, March T. Ma and M. Deisher, "Search Techniques in Speech Recognition," Intel internal technical report, September T. Ma, "Linear Dynamic Models (LDM) for Automatic Speech Recognition," Intel Intern Seminar Series, August 2008.

Slide 23 References [1]Lawrence R. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, Readings in speech recognition, Morgan Kaufmann Publishers Inc., San Francisco, CA, 1990 [2]L.R. Rabiner and B.H. Juang, Fundamentals of Speech Recognition, Prentice Hall, Englewood Cliffs, New Jersey, USA, [3]J. Picone, “Continuous Speech Recognition Using Hidden Markov Models,” IEEE Acoustics, Speech, and Signal Processing Magazine, vol. 7, no. 3, pp , July [4]Digalakis, V., Rohlicek, J. and Ostendorf, M., “ML Estimation of a Stochastic Linear System with the EM Algorithm and Its Application to Speech Recognition,” IEEE Transactions on Speech and Audio Processing, vol. 1, no. 4, pp. 431–442, October [5]Frankel, J. and King, S., “Speech Recognition Using Linear Dynamic Models,” IEEE Transactions on Speech and Audio Processing, vol. 15, no. 1, pp. 246–256, January [6]S. Renals, Speech and Neural Network Dynamics, Ph. D. dissertation, University of Edinburgh, UK, 1990 [7]J. Tebelskis, Speech Recognition using Neural Networks, Ph. D. dissertation, Carnegie Mellon University, Pittsburg, USA, 1995 [8]A. Ganapathiraju, J. Hamaker and J. Picone, "Applications of Support Vector Machines to Speech Recognition," IEEE Transactions on Signal Processing, vol. 52, no. 8, pp , August [9]J. Hamaker and J. Picone, "Advances in Speech Recognition Using Sparse Bayesian Methods," submitted to the IEEE Transactions on Speech and Audio Processing, January 2003.

Slide 24 Thank you! Questions?