Midterm Review. The Midterm Everything we have talked about so far Stuff from HW I won’t ask you to do as complicated calculations as the HW Don’t need.

Slides:



Advertisements
Similar presentations
Lecture 16 Hidden Markov Models. HMM Until now we only considered IID data. Some data are of sequential nature, i.e. have correlations have time. Example:
Advertisements

Basics of Statistical Estimation
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Image Modeling & Segmentation
Mixture Models and the EM Algorithm
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Learning HMM parameters
Expectation Maximization
Dynamic Bayesian Networks (DBNs)
Supervised Learning Recap
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 143, Brown James Hays 02/22/11 Many slides from Derek Hoiem.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 03/15/12.
Graphical models: approximate inference and learning CA6b, lecture 5.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
… Hidden Markov Models Markov assumption: Transition model:
PatReco: Hidden Markov Models Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
First introduced in 1977 Lots of mathematical derivation Problem : given a set of data (data is incomplete or having missing values). Goal : assume the.
Ch 13. Sequential Data (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Kim Jin-young Biointelligence Laboratory, Seoul.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
Part 4 b Forward-Backward Algorithm & Viterbi Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
. Hidden Markov Models Lecture #5 Prepared by Dan Geiger. Background Readings: Chapter 3 in the text book (Durbin et al.).
Lecture 5: Learning models using EM
Most slides from Expectation Maximization (EM) Northwestern University EECS 395/495 Special Topics in Machine Learning.
. Hidden Markov Models For Genetic Linkage Analysis Lecture #4 Prepared by Dan Geiger.
Part 4 c Baum-Welch Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Maximum Likelihood (ML), Expectation Maximization (EM)
. Learning Parameters of Hidden Markov Models Prepared by Dan Geiger.
Expectation-Maximization
Bayesian Networks Alan Ritter.
Expectation-Maximization (EM) Chapter 3 (Duda et al.) – Section 3.9
Learning HMM parameters Sushmita Roy BMI/CS 576 Oct 21 st, 2014.
Gaussian Mixture Models and Expectation Maximization.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Computer vision: models, learning and inference
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
1 Naïve Bayes Models for Probability Estimation Daniel Lowd University of Washington (Joint work with Pedro Domingos)
EM and expected complete log-likelihood Mixture of Experts
Bayesian Inference Ekaterina Lomakina TNU seminar: Bayesian inference 1 March 2013.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Lecture 19: More EM Machine Learning April 15, 2010.
Statistical Learning (From data to distributions).
Probability and Measure September 2, Nonparametric Bayesian Fundamental Problem: Estimating Distribution from a collection of Data E. ( X a distribution-valued.
Lecture 17 Gaussian Mixture Models and Expectation Maximization
CS Statistical Machine learning Lecture 24
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
Lecture 2: Statistical learning primer for biologists
Learning Sequence Motifs Using Expectation Maximization (EM) and Gibbs Sampling BMI/CS 776 Mark Craven
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
RADFORD M. NEAL GEOFFREY E. HINTON 발표: 황규백
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/22/11.
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Bayesian Belief Propagation for Image Understanding David Rosenberg.
Information Bottleneck versus Maximum Likelihood Felix Polyakov.
Today.
CS 2750: Machine Learning Density Estimation
Statistical Models for Automatic Speech Recognition
More about Posterior Distributions
Bayesian Models in Machine Learning
Statistical Models for Automatic Speech Recognition
Expectation Maximization
LECTURE 07: BAYESIAN ESTIMATION
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

Midterm Review

The Midterm Everything we have talked about so far Stuff from HW I won’t ask you to do as complicated calculations as the HW Don’t need a calculator No books / notes

Maximum Likelihood Estimation How to apply the maximum likelihood principle – log likelihood + derivative + solve for 0 – You should know how to do this for Bernoulli trials and 1-D Gaussian Conjugate distributions – Dirichlet, Beta

Mixture Models and EM What does the EM algorithm do? – Understand the E-step and M-step Log-exp-sum trick – You should be able to derive this – You should understand why we need to use it

Hidden Markov Models Viterbi – What does it do? – What is the running time? Forward-backward – What does it do? Be able to compute the probability of a “parse” – Joint probability of a sequence of observed and hidden states

Bayesian Networks Understand d-separation criteria Be able to answer simple questions about whether variables are independent given some evidence Markov Blanket

Markov Networks / Belief Propagation Moralizing a graph (convert Bayesian network into Markov Network) Belief propagation – What does it do, when is it guaranteed to converge to the correct posterior distribution.