K-means and Gaussian Mixture Model 王养浩 2013 年 11 月 20 日.

Slides:



Advertisements
Similar presentations
Part 2: Unsupervised Learning
Advertisements

Image Modeling & Segmentation
Hierarchical Clustering, DBSCAN The EM Algorithm
Mixture Models and the EM Algorithm
Unsupervised Learning
Expectation Maximization
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Supervised Learning Recap
The EM algorithm LING 572 Fei Xia Week 10: 03/09/2010.
Intro to DPM By Zhangliliang. Outline Intuition Introduction to DPM Model Inference(matching) Training latent SVM Training Procedure Initialization Post-processing.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
K-means clustering Hongning Wang
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
First introduced in 1977 Lots of mathematical derivation Problem : given a set of data (data is incomplete or having missing values). Goal : assume the.
Most slides from Expectation Maximization (EM) Northwestern University EECS 395/495 Special Topics in Machine Learning.
Gaussian Mixture Example: Start After First Iteration.
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Modeling of Mel Frequency Features for Non Stationary Noise I.AndrianakisP.R.White Signal Processing and Control Group Institute of Sound and Vibration.
Expectation-Maximization
Visual Recognition Tutorial
Part 3 Vector Quantization and Mixture Density Model CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Clustering with Bregman Divergences Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh Presented by Rohit Gupta CSci 8980: Machine Learning.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
EM algorithm LING 572 Fei Xia 03/02/06. Outline The EM algorithm EM for PM models Three special cases –Inside-outside algorithm –Forward-backward algorithm.
Gaussian Mixture Models and Expectation Maximization.
Incomplete Graphical Models Nan Hu. Outline Motivation K-means clustering Coordinate Descending algorithm Density estimation EM on unconditional mixture.
Math 5364 Notes Chapter 8: Cluster Analysis Jesse Crawford Department of Mathematics Tarleton State University.
Lecture 19: More EM Machine Learning April 15, 2010.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
The Group Lasso for Logistic Regression Lukas Meier, Sara van de Geer and Peter Bühlmann Presenter: Lu Ren ECE Dept., Duke University Sept. 19, 2008.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –Unsupervised Learning: Kmeans, GMM, EM Readings: Barber
Lecture 17 Gaussian Mixture Models and Expectation Maximization
Lecture 6 Spring 2010 Dr. Jianjun Hu CSCE883 Machine Learning.
CS Statistical Machine learning Lecture 24
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Lecture 2: Statistical learning primer for biologists
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Clustering (1) Chapter 7. Outline Introduction Clustering Strategies The Curse of Dimensionality Hierarchical k-means.
CSE 446: Expectation Maximization (EM) Winter 2012 Daniel Weld Slides adapted from Carlos Guestrin, Dan Klein & Luke Zettlemoyer.
Advanced Artificial Intelligence Lecture 8: Advance machine learning.
1 Cluster Analysis – 2 Approaches K-Means (traditional) Latent Class Analysis (new) by Jay Magidson, Statistical Innovations based in part on a presentation.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –(Finish) Expectation Maximization –Principal Component Analysis (PCA) Readings:
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
Expectation-Maximization (EM)
Lecture 18 Expectation Maximization
LECTURE 10: EXPECTATION MAXIMIZATION (EM)
10701 / Machine Learning.
Latent Variables, Mixture Models and EM
Probabilistic Models for Linear Regression
K-means and Hierarchical Clustering
Expectation-Maximization
TOP DM 10 Algorithms C4.5 C 4.5 Research Issue:
Probabilistic Models with Latent Variables
Expectation Maximization
Stochastic Optimization Maximization for Latent Variable Models
INTRODUCTION TO Machine Learning
10701 Recitation Pengtao Xie
Lecture 11 Generalizations of EM.
Machine Learning – a Probabilistic Perspective
EM Algorithm 主講人:虞台文.
Criterion Methods and Approximation Concepts
Clustering (2) & EM algorithm
Introduction to Machine learning
Stochastic Methods.
Presentation transcript:

K-means and Gaussian Mixture Model 王养浩 2013 年 11 月 20 日

Outline K-means Gaussian Mixture Model Expectation Maximum

K-means Gather data points to a few cohesive ‘Clusters’ Unsupervised Learning

K-means

Easy Fast Euclidean distance? K needs input ? Convergence ?

Determination of K Rule of Thumb : Elbow Method Cross Validation

K-means Convergence x (i) data points μ c(i) cluster centroids Coordinate descent

Coordinate Descent

K-means Convergence Non-circle Clusters

K-means Convergence Local minimum – The optimization object is non-convex

Gaussian Mixture Model Mixture of Gaussian distribution

Gaussian Mixture Model Log likelihood Maximum likelihood – Expectation Maximum

Expectation Maximum

Jenson inquality

Expectation Maximum

Construct lower bound

Expectation Maximum

Repeat until convergence

Generalized Expectation Maximum Difficulty in M-step

Summary K-means – Coordinate descent Gaussian Mixture Model – Expectation Maximum Expectation Maximum – MLE for models with latent variables – Generalized EM

Thanks!