Recitation on EM slides taken from:

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Expectation Maximization Dekang Lin Department of Computing Science University of Alberta.
Image Modeling & Segmentation
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Confidence Intervals This chapter presents the beginning of inferential statistics. We introduce methods for estimating values of these important population.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Part of Speech Tagging The DT students NN went VB to P class NN Plays VB NN well ADV NN with P others NN DT Fruit NN flies NN VB NN VB like VB P VB a DT.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
. Learning – EM in ABO locus Tutorial #08 © Ydo Wexler & Dan Geiger.
. Learning – EM in The ABO locus Tutorial #8 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
 CpG is a pair of nucleotides C and G, appearing successively, in this order, along one DNA strand.  CpG islands are particular short subsequences in.
Overview Full Bayesian Learning MAP learning
First introduced in 1977 Lots of mathematical derivation Problem : given a set of data (data is incomplete or having missing values). Goal : assume the.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
Hidden Markov Model Special case of Dynamic Bayesian network Single (hidden) state variable Single (observed) observation variable Transition probability.
1 An Introduction to Statistical Machine Translation Dept. of CSIE, NCKU Yao-Sheng Chang Date:
Lecture 5: Learning models using EM
Most slides from Expectation Maximization (EM) Northwestern University EECS 395/495 Special Topics in Machine Learning.
CpG islands in DNA sequences
. Learning – EM in The ABO locus Tutorial #9 © Ilan Gronau.
Maximum likelihood (ML) and likelihood ratio (LR) test
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference (Sec. )
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Clustering.
Expectation Maximization Algorithm
Expectation-Maximization
Computer vision: models, learning and inference
EM Algorithm Likelihood, Mixture Models and Clustering.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
. Class 5: Hidden Markov Models. Sequence Models u So far we examined several probabilistic model sequence models u These model, however, assumed that.
. Learning – EM in The ABO locus Tutorial #9 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
EM algorithm LING 572 Fei Xia 03/02/06. Outline The EM algorithm EM for PM models Three special cases –Inside-outside algorithm –Forward-backward algorithm.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference.
Computer vision: models, learning and inference
Likelihood probability of observing the data given a model with certain parameters Maximum Likelihood Estimation (MLE) –find the parameter combination.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Maximum likelihood estimation of relative transcript abundances Advanced bioinformatics 2012.
AP Statistics Chapter 9 Notes.
Model Inference and Averaging
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Comp. Genomics Recitation 12 Bayesian networks Taken from Artificial Intelligence course, MIT, 6.034
Motif finding with Gibbs sampling CS 466 Saurabh Sinha.
P ROBABLITY S TATICS &. PROJECT. 1 Assuming that the error terms are distributed as: Please derive the maximum likelihood estimator for the simple linear.
Lecture 13: Linkage Analysis VI Date: 10/08/02  Complex models  Pedigrees  Elston-Stewart Algorithm  Lander-Green Algorithm.
HMM - Part 2 The EM algorithm Continuous density HMM.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Lecture 2: Statistical learning primer for biologists
Sampling and estimation Petter Mostad
Flat clustering approaches
Learning Sequence Motifs Using Expectation Maximization (EM) and Gibbs Sampling BMI/CS 776 Mark Craven
CHAPTER 9 Inference: Estimation The essential nature of inferential statistics, as verses descriptive statistics is one of knowledge. In descriptive statistics,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Review of statistical modeling and probability theory Alan Moses ML4bio.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Other Models for Time Series. The Hidden Markov Model (HMM)
. The EM algorithm Lecture #11 Acknowledgement: Some slides of this lecture are due to Nir Friedman.
MathematicalMarketing Slide 3c.1 Mathematical Tools Chapter 3: Part c – Parameter Estimation We will be discussing  Nonlinear Parameter Estimation  Maximum.
Classification of unlabeled data:
Learning Sequence Motif Models Using Expectation Maximization (EM)
Latent Variables, Mixture Models and EM
More about Posterior Distributions
10701 / Machine Learning Today: - Cross validation,
EM for Inference in MV Data
EM for Inference in MV Data
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

Recitation on EM slides taken from: Computational Genomics Recitation #6

All EM questions are in the format: 1.Write the likelihood function. 2.Write the Q function. 3.Derive the update rule.

Estimation problems

What is the unobserved data in this case?

Estimation problems

? ? ?

? ? ?

? ? ? ? ? ? ? ? ?

? ? ?

EM question Let G = (G 1, …, G n ) be n contiguous DNA regions representing genes. For each G i we define the mRNA concentration of the gene as P i, s.t. their sum is equal to 1. P = (P 1, …, P n ) can be interpreted as the normalized expression levels for the regions in G.

EM question Our model assumes that reads are generated by randomly picking a region R from G according to the distribution P, and then copying this region. The copying process is error-prone. This process is repeated until we have a set of m reads R = r 1, …, r m generated according to the model described above.

EM question For each region G j and read r i, we have a probability p ij = P(r j | G i ), the probability of observing r j given that the locus of the read was gene G i. In practice, for each read r j, this probability will be close to zero for all but a few regions.

Likelihood function Write the likelihood of observing the m reads. ?

Q function Write the Q(P | P (t) ) term. ? ?

M-step Write the M-step term using argmax function.

Update rule Infer from c the update step for P. When we want to maximize ∑ i a i log(P i ) based on P i, we achieve the maximum P i =a i /∑ i a i ?