Bayesian Models for Gene expression With DNA Microarray Data Joseph G. Ibrahim, Ming-Hui Chen, and Robert J. Gray Presented by Yong Zhang.

Slides:



Advertisements
Similar presentations
MCMC estimation in MlwiN
Advertisements

Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
1 Parametric Empirical Bayes Methods for Microarrays 3/7/2011 Copyright © 2011 Dan Nettleton.
Bayesian dynamic modeling of latent trait distributions Duke University Machine Learning Group Presented by Kai Ni Jan. 25, 2007 Paper by David B. Dunson,
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Lecture XXIII.  In general there are two kinds of hypotheses: one concerns the form of the probability distribution (i.e. is the random variable normally.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
K Means Clustering , Nearest Cluster and Gaussian Mixture
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Visual Recognition Tutorial
A Bayesian  2 test for goodness of fit 10/23/09 Multilevel RIT.
L15:Microarray analysis (Classification) The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Continuous Random Variables and Probability Distributions
Lecture 5: Learning models using EM
Visual Recognition Tutorial
Visual Recognition Tutorial
Linear and generalised linear models
Computer vision: models, learning and inference
Linear and generalised linear models
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Multivariate Data and Matrix Algebra Review BMTRY 726 Spring 2012.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Bayesian Model Selection in Factorial Designs Seminal work is by Box and Meyer Seminal work is by Box and Meyer Intuitive formulation and analytical approach,
Pattern Recognition: Baysian Decision Theory Charles Tappert Seidenberg School of CSIS, Pace University.
Principles of Pattern Recognition
1 A Presentation of ‘Bayesian Models for Gene Expression With DNA Microarray Data’ by Ibrahim, Chen, and Gray Presentation By Lara DePadilla.
On Model Validation Techniques Alex Karagrigoriou University of Cyprus "Quality - Theory and Practice”, ORT Braude College of Engineering, Karmiel, May.
Continuous Random Variables
A Beginner’s Guide to Bayesian Modelling Peter England, PhD EMB GIRO 2002.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Computer Science, Software Engineering & Robotics Workshop, FGCU, April 27-28, 2012 Fault Prediction with Particle Filters by David Hatfield mentors: Dr.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
Probability and Measure September 2, Nonparametric Bayesian Fundamental Problem: Estimating Distribution from a collection of Data E. ( X a distribution-valued.
Bayesian Multivariate Logistic Regression by Sean O’Brien and David Dunson (Biometrics, 2004 ) Presented by Lihan He ECE, Duke University May 16, 2008.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
HMM - Part 2 The EM algorithm Continuous density HMM.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
BCS547 Neural Decoding.
Bayesian Decision Theory Basic Concepts Discriminant Functions The Normal Density ROC Curves.
A shared random effects transition model for longitudinal count data with informative missingness Jinhui Li Joint work with Yingnian Wu, Xiaowei Yang.
Item Parameter Estimation: Does WinBUGS Do Better Than BILOG-MG?
The Variance of a Random Variable Lecture 35 Section Fri, Mar 26, 2004.
Javad Azimi, Ali Jalali, Xiaoli Fern Oregon State University University of Texas at Austin In NIPS 2011, Workshop in Bayesian optimization, experimental.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Univariate Gaussian Case (Cont.)
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Univariate Gaussian Case (Cont.)
Lecture 1.31 Criteria for optimal reception of radio signals.
Bayesian Semi-Parametric Multiple Shrinkage
Parameter Estimation 主講人:虞台文.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
POINT ESTIMATOR OF PARAMETERS
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Reversing Label Switching:
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Mean and Standard Deviation
Parametric Empirical Bayes Methods for Microarrays
LECTURE 09: BAYESIAN LEARNING
LECTURE 07: BAYESIAN ESTIMATION
6.3 Sampling Distributions
Yalchin Efendiev Texas A&M University
Classical regression review
Presentation transcript:

Bayesian Models for Gene expression With DNA Microarray Data Joseph G. Ibrahim, Ming-Hui Chen, and Robert J. Gray Presented by Yong Zhang

Goals: 1)To build a model to compare between normal and tumor tissues and to find the genes that best distinguish between tissue types. 2)to develop model assessment techniques so as to assess the fit of a class of competing models.

Outline General Model Gene Selection Algo. Prior Distributions L measures(assessment) example

Data structure x: the expression level for a given gene C 0 : threshold value for which a gene is considered as not expressed Let p = P(x=c 0 ), then where y is the continuous part for x.

j=1, 2 index the tissue type(normal vs. tumor) i=1,2,…n j, ith individual g=1,…G, gth gene x jig : the gene expression mixture random variable for the jth tissue type for the ith individual and the gth gene.

The General Model Assume δ jig = 1(x jig =c 0 ) p jg =P(x jig =c 0 )=P(δ jig = 1)

 =( ,  2,p) Data D=(x 111,…x 2, n2,G,  ) Likelihood function for  : L(  |D)= In order to find which genes best discriminate between the normal and tumor tissues, let

Then we set such that we can use  g to judge them.

Prior Distributions  jg 2 ~ Inverse Gamma(a j0,b j0 )  j0 ~ N(m j0,v j0 2 ), j=1,2

b j0 ~ gamma(q j0,t j0 ) e jg ~ N(u j0,k j0 w j0 2 )

Gene Selection Algo. 1)For each gene, compute  g and 2)Select a “threshold” value, say r 0, to decide which genes are different. If 3)Once the gth genes are declared different, set  1g   2g, otherwise set  1g =  2g   g, where  g is treated as unknown.

Gene Selection Algo. 4) Create several submodels using several values of r 0. 5) Use L measure to decide which submodel is the best one(smallest L measure).

The properties of this approach 1)Model the gene expression level as a mixture random variable. 2)Use a lognormal model for the continuous part of the mixture. 3)Use L measure statistic for evaluating models.

L measure for model assessment It relies on the notion of an imaginary replicate experiment. Let z= (z 111, …, z 2,n2,G ) denote future values of a replicate experiment.

L measure is the expected squared Euclidean distance between x and z, A more general is The r.s. of the last formula can be got by MCMC.

Computational Algo.(MCMC) 1.

For 1–4 and 6, the generation is straightforward. For 5, we can use an adaptive rejection algorithm(Gilks and Wild, 1992) because the corresponding conditional posterior densities are log-concave.

Discussion That model development and prior distributions in this paper can be easily extended to handle three or more tissue types. More general classes of priors The gene selection criterions