1 Expectation Maximization Algorithm José M. Bioucas-Dias Instituto Superior Técnico 2005.

Slides:



Advertisements
Similar presentations
Image Modeling & Segmentation
Advertisements

Expectation-Maximization (EM) Algorithm Md. Rezaul Karim Professor Department of Statistics University of Rajshahi Bangladesh September 21, 2012.
EMNLP, June 2001Ted Pedersen - EM Panel1 A Gentle Introduction to the EM Algorithm Ted Pedersen Department of Computer Science University of Minnesota.
An Introduction to the EM Algorithm Naala Brewer and Kehinde Salau.
Unsupervised Learning
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
The EM algorithm LING 572 Fei Xia Week 10: 03/09/2010.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
. Parameter Estimation and Relative Entropy Lecture #8 Background Readings: Chapters 3.3, 11.2 in the text book, Biological Sequence Analysis, Durbin et.
First introduced in 1977 Lots of mathematical derivation Problem : given a set of data (data is incomplete or having missing values). Goal : assume the.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
5. CONCLUSIONS Initially constructed and applied for the New York State’s rainfall stations (Wilks, 1998), DUM was able to reproduce its good performance.
The EM algorithm (Part 1) LING 572 Fei Xia 02/23/06.
Parametric Inference.
EM Algorithm Presented By: Haiguang Li Computer Science Department University of Vermont Fall 2011.
Expectation-Maximization
Learning and Vision Seminar Anand D. Subramaniam Seminar on Vision and Learning University of California, San Diego September 20, 2001 Learning and Recognizing.
What is it? When would you use it? Why does it work? How do you implement it? Where does it stand in relation to other methods? EM algorithm reading group.
Improved characterization of neural and behavioral response properties using point-process state-space framework Anna Alexandra Dreyer Harvard-MIT Division.
Expectation-Maximization (EM) Chapter 3 (Duda et al.) – Section 3.9
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
TRICAP_06 Mohamed Hanafi PLS PATH MODELLING : Computation of latent variables with the estimation mode B UNITE DE SENSOMETRIE ET CHIMIOMETRIE Nantes-France.
© 2007 John M. Abowd, Lars Vilhuber, all rights reserved Estimating m and u Probabilities Using EM Based on Winkler 1988 "Using the EM Algorithm for Weight.
ADVANCED SIGNAL PROCESSING TECHNIQUES FOR WIRELESS COMMUNICATIONS Erdal Panayırcı Electronics Engineering Department IŞIK University.
Statistical Methods for Missing Data Roberta Harnett MAR 550 October 30, 2007.
Case Study – Gumbel Distribution
Math 5364 Notes Chapter 8: Cluster Analysis Jesse Crawford Department of Mathematics Tarleton State University.
1 HMM - Part 2 Review of the last lecture The EM algorithm Continuous density HMM.
November 1, 2012 Presented by Marwan M. Alkhweldi Co-authors Natalia A. Schmid and Matthew C. Valenti Distributed Estimation of a Parametric Field Using.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Eurostat Statistical matching when samples are drawn according to complex survey designs Training Course «Statistical Matching» Rome, 6-8 November 2013.
An Introduction to the EM Algorithm By Naala Brewer and Kehinde Salau Project Advisor – Prof. Randy Eubank Advisor – Prof. Carlos Castillo-Chavez MTBI,
HMM - Part 2 The EM algorithm Continuous density HMM.
Machine Learning Saarland University, SS 2007 Holger Bast Max-Planck-Institut für Informatik Saarbrücken, Germany Lecture 8, Friday June 8 th, 2007 (introduction.
ECE 8443 – Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem Proof EM Example – Missing Data Intro to Hidden Markov Models.
Expectation-Maximization (EM) Algorithm & Monte Carlo Sampling for Inference and Approximation.
A shared random effects transition model for longitudinal count data with informative missingness Jinhui Li Joint work with Yingnian Wu, Xiaowei Yang.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition Objectives: Reestimation Equations Continuous Distributions Gaussian Mixture Models EM Derivation of Reestimation Resources:
1 Cluster Analysis – 2 Approaches K-Means (traditional) Latent Class Analysis (new) by Jay Magidson, Statistical Innovations based in part on a presentation.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Maximum Likelihood Estimates and the EM Algorithms III Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University
For multivariate data of a continuous nature, attention has focussed on the use of multivariate normal components because of their computational convenience.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Graduate : Yu Cheng Chen Author: Lynette.
Markov-Chain-Monte-Carlo (MCMC) & The Metropolis-Hastings Algorithm P548: Intro Bayesian Stats with Psych Applications Instructor: John Miyamoto 01/19/2016:
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
30 November, 2005 CIMCA2005, Vienna 1 Statistical Learning Procedure in Loopy Belief Propagation for Probabilistic Image Processing Kazuyuki Tanaka Graduate.
10 October, 2007 University of Glasgow 1 EM Algorithm with Markov Chain Monte Carlo Method for Bayesian Image Analysis Kazuyuki Tanaka Graduate School.
ECE 7251: Signal Detection and Estimation
LECTURE 10: EXPECTATION MAXIMIZATION (EM)
Graduate School of Information Sciences, Tohoku University, Japan
Expectation-Maximization
دانشگاه صنعتی امیرکبیر Instructor : Saeed Shiry
Bayesian Models in Machine Learning
Expectation Maximization Mixture Models HMMs
Introduction to EM algorithm
'Linear Hierarchical Models'
信道匹配算法用于混合模型——挑战EM算法 Channels’ Matching Algorithm for Mixture Models ——A Challenge to the EM Algorithm 鲁晨光 Chenguang Lu Hpage:
EM for Inference in MV Data
Stat 223 Introduction to the Theory of Statistics
Probabilistic image processing and Bayesian network
Probabilistic image processing and Bayesian network
From the EM Algorithm to the CM-EM Algorithm for Global Convergence of Mixture Models Chenguang Lu 从EM算法到CM-EM算法求混合模型全局收敛.
Lecture 11 Generalizations of EM.
EM for Inference in MV Data
Clustering (2) & EM algorithm
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

1 Expectation Maximization Algorithm José M. Bioucas-Dias Instituto Superior Técnico 2005

2 Expectation Maximization (EM) Tool EM approach: Approximate by a tractable iterative procedure Problem: Compute the MAP estimate is often very hard to compute [Dempster et. al, 77], [Little & Rubin, 87], [McLachlan and T. Krishnan, 97]

3 Complete and Missing Data Let where h is a non-invertible function such that Particular case: are termed missing data and complete data, respectively

4 A Minorization Bound for Define E is the Kullback-Leibler distance between the densities Note: Given two densities p and q Facts:

5 EM Concept

6 EM Algorithm Define the sequence Then is non-decreasing Proof Kullback Maximization

7 EM Algorithm (Cont.) Computing Q It is not necessary to compute the second mean value (it does not depend on )

8 EM Acronym E – Expectation M – Maximization

9 The EM rationale remains valid if we replace with i.e., it is not neccessary to maximaze w.r.t. It suffices to assure that increases. Generalized EM (GEM) Algorithm

10 Generalized EM (GEM) Algorithm (cont.)

11 Convergency of Assume that is continuous on both arguments then all limit points of are stationary points of and the GEM sequence converges monotonically to for some stationary point [Wu, 83].

12 References A. Dempster, N. Laird, and D. Rubin. “Maximum likelihood estimation from incomplete data via the EM algorithm.” Journal of the Royal Statistical Society B, vol. 39, pp. 1-38, R. Little and D. Rubin. Statistical Analysis with Missing Data. John Wiley & Sons, New York, 1987 G. McLachlan and T. Krishnan. The EM Algorithm and Extensions. John Wiley & Sons, New York, M. Tanner. Tools for Statistical Inference. Springer-Verlag, New York, C. Wu, “On the convergence properties of the EM algorithm,” The Annal of Statistics, vol. 11, no. 1, pp , 1983