Inter Class MLLR for Speaker Adaptation

Slides:



Advertisements
Similar presentations
Kriging.
Advertisements

Economics 20 - Prof. Anderson1 Testing for Unit Roots Consider an AR(1): y t =  +  y t-1 + e t Let H 0 :  = 1, (assume there is a unit root) Define.
A posteriori Error Estimate - Adaptive method Consider the boundary value problem Weak form Discrete Equation Error bounds ( priori error )
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Single-Channel Speech Enhancement in Both White and Colored Noise Xin Lei Xiao Li Han Yan June 5, 2002.
Speaker Adaptation in Sphinx 3.x and CALO David Huggins-Daines
Speaker Adaptation for Vowel Classification
Chapter 4 Multiple Regression.
1Prof. Dr. Rainer Stachuletz Testing for Unit Roots Consider an AR(1): y t =  +  y t-1 + e t Let H 0 :  = 1, (assume there is a unit root) Define 
Chapter 4 Multiple Regression. 4.1 Introduction.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
Optimal Adaptation for Statistical Classifiers Xiao Li.
1 USING CLASS WEIGHTING IN INTER-CLASS MLLR Sam-Joo Doh and Richard M. Stern Department of Electrical and Computer Engineering and School of Computer Science.
Bootstrap Estimation of the Predictive Distributions of Reserves Using Paid and Incurred Claims Huijuan Liu Cass Business School Lloyd’s of London 10/07/2007.
Introduction to Linear Mixed Effects Kiran Pedada PhD Student (Marketing) March 26, 2015.
Adaptive autocalibration of unknown transducer positions in time reversal imaging Raghuram Rangarajan EE Systems University Of Michigan, Ann Arbor.
CSE 330: Numerical Methods.  Regression analysis gives information on the relationship between a response (dependent) variable and one or more predictor.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Linear Trend Lines = b 0 + b 1 X t Where is the dependent variable being forecasted X t is the independent variable being used to explain Y. In Linear.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
1 Dr. Jerrell T. Stracener EMIS 7370 STAT 5340 Probability and Statistics for Scientists and Engineers Department of Engineering Management, Information.
MTH 161: Introduction To Statistics
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Brian Macpherson Ph.D, Professor of Statistics, University of Manitoba Tom Bingham Statistician, The Boeing Company.
Statistical analysis Outline that error bars are a graphical representation of the variability of data. The knowledge that any individual measurement.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Impulse Burst Position Detection and Channel Estimation Scheme for OFDM Systems 高永安 老師 學生:吳林達 報告日期: 2006/5/18.
1 Introduction What does it mean when there is a strong positive correlation between x and y ? Regression analysis aims to find a precise formula to relate.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Regression. We have talked about regression problems before, as the problem of estimating the mapping f(x) between an independent variable x and a dependent.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
A comparative approach for gene network inference using time-series gene expression data Guillaume Bourque* and David Sankoff *Centre de Recherches Mathématiques,
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Computacion Inteligente Least-Square Methods for System Identification.
Flexible Speaker Adaptation using Maximum Likelihood Linear Regression Authors: C. J. Leggetter P. C. Woodland Presenter: 陳亮宇 Proc. ARPA Spoken Language.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
CSE 330: Numerical Methods. What is regression analysis? Regression analysis gives information on the relationship between a response (dependent) variable.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
MathematicalMarketing Slide 3c.1 Mathematical Tools Chapter 3: Part c – Parameter Estimation We will be discussing  Nonlinear Parameter Estimation  Maximum.
Multiple Regression.
Ch5 Relaxing the Assumptions of the Classical Model
Statistical analysis.
12. Principles of Parameter Estimation
(5) Notes on the Least Squares Estimate
Basic Estimation Techniques
Statistical analysis.
Inference: Conclusion with Confidence
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Linear Regression (continued)
Simple Linear Regression - Introduction
A special case of calibration
Making Statistical Inferences
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
CHAPTER 29: Multiple Regression*
Unfolding Problem: A Machine Learning Approach
Multiple Regression.
Software life cycle models
Regression Models - Introduction
Model Comparison: some basic concepts
10701 / Machine Learning Today: - Cross validation,
BASIC REGRESSION CONCEPTS
Nonlinear Fitting.
Basic Practice of Statistics - 3rd Edition Introduction to Inference
Linear Panel Data Models
12. Principles of Parameter Estimation
Calibration and homographies
Regression and Correlation of Data
Regression Models - Introduction
Presentation transcript:

Inter Class MLLR for Speaker Adaptation Presenter : 陳彥達

Reference Sam-Joo Doh and Richard M. Stern, “Inter-Class MLLR for Speaker Adaptation”, ICASSP 2000.

Outline Introduction Concept of inter-class MLLR Inter-class function training Experiments

Introduction Why do we use MLLR ? Simple idea Few unknown parameters Sparse adaptation data Rapid adaptation

Introduction(2) i all Gaussian , i Class 1 i Class 2

Introduction(3) Single-class MLLR Multi-class MLLR All parameters use the same transformation in adaptation More reliable for small amount of adaptation data Multi-class MLLR Parameters in different classes use different transformation in adaptation More reliable for large amount of adaptation data

Introduction(4) Shortcoming of conventional MLLR The number of classes should be carried out according to the amount of adaptation data. Classes are independent in multi-class MLLR, so some parameters may not be adapted. Main idea of inter-class MLLR Use correlation between different classes to compensate for the shortcoming mentioned above.

Concept of inter-class MLLR Inter-class function The relation between different classes if is the inter-class function between class 1 and class 2 , i class 1 , i class 2

Concept of inter-class MLLR(2) Model setup steps Define multiple classes ( say class 1~n ) Find inter-class functions between each class Adaptation steps Choose a target class which is going to be adapted ( say class k ) Rank other classes according to their “closeness” to the target class

Concept of inter-class MLLR(3) If adaptation data ( i target class k ) coming, use conventional MLLR to find If adaptation data ( i target class k ) coming, use inter-class function to convert as the adaptation data in class k, and then use conventional MLLR to find Repeat above steps until all classes are adapted

Concept of inter-class MLLR(4) Adaptation data are selected from classes of decreasing proximity to the target class until there are sufficient data to estimate the target function. Limit cases no neighboring classes used → conventional multi-class MLLR All neighboring classes used → conventional single-class MLLR

Inter-class function training , i class m , i class n let then , i class n

Inter-class function training(2) Assuming we have training data of R speakers. We use these data to train , for each class for each speaker. ie. s={1,2,…,R}, m={1,2,…,n}. , i class n, for class m for speaker s ∴

Inter-class function training(3) Let we use the equation above and training data for all speakers to obtain and by conventional MLLR.

Experiments Mean-square error from simulated estimates of Gaussian means

Experiments(2) Word error rate for different types of MLLR 25 training speakers 13 phonetic-based regression classes 10 testing speakers, 20 sentences per speaker for testing, 5 sentences per speaker for adaptation use all the neighbor classes to estimate each target class Silence and noise phones are not adapted

Experiments(3)