Rainer Spang, Max Planck Institute for Molecular Genetics, Berlin

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Linear Models for Microarray Data
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
The Normal Distribution
INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Pattern Recognition and Machine Learning
Bayesian Factor Regression Models in the “Large p, Small n” Paradigm Mike West, Duke University Presented by: John Paisley Duke University.
Computer vision: models, learning and inference Chapter 8 Regression.
Achim Tresch Computational Biology ‘Omics’ - Analysis of high dimensional Data.
Dimension reduction (1)
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Industrial Engineering College of Engineering Bayesian Kernel Methods for Binary Classification and Online Learning Problems Theodore Trafalis Workshop.
Computational Diagnostics We are a new research group in the department of Computational Molecular Biology at the Max Planck Institute for Molecular Genetics.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Predictive sub-typing of subjects Retrospective and prospective studies Exploration of clinico-genomic data Identify relevant gene expression patterns.
Linear and generalised linear models
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Study of Sparse Online Gaussian Process for Regression EE645 Final Project May 2005 Eric Saint Georges.
Gene Expression Based Tumor Classification Using Biologically Informed Models ISI 2003 Berlin Claudio Lottaz und Rainer Spang Computational Diagnostics.
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
Gaussian Processes for Transcription Factor Protein Inference Neil D. Lawrence, Guido Sanguinetti and Magnus Rattray.
Model Assessment and Selection Florian Markowetz & Rainer Spang Courses in Practical DNA Microarray Analysis.
Bayes Factor Based on Han and Carlin (2001, JASA).
PATTERN RECOGNITION AND MACHINE LEARNING
Classification (Supervised Clustering) Naomi Altman Nov '06.
The horseshoe estimator for sparse signals CARLOS M. CARVALHO NICHOLAS G. POLSON JAMES G. SCOTT Biometrika (2010) Presented by Eric Wang 10/14/2010.
Molecular Diagnosis Florian Markowetz & Rainer Spang Courses in Practical DNA Microarray Analysis.
ArrayCluster: an analytic tool for clustering, data visualization and module finder on gene expression profiles 組員:李祥豪 謝紹陽 江建霖.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Bayesian Analysis and Applications of A Cure Rate Model.
Sample classification using Microarray Data. AB We have two sample entities malignant vs. benign tumor patient responding to drug vs. patient resistant.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Simulation of the matrix Bingham-von Mises- Fisher distribution, with applications to multivariate and relational data Discussion led by Chunping Wang.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Computational Diagnostics A new research group at the Max Planck Institute for molecular Genetics, Berlin.
Paper Reading Dalong Du Nov.27, Papers Leon Gu and Takeo Kanade. A Generative Shape Regularization Model for Robust Face Alignment. ECCV08. Yan.
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
Introduction to Machine Learning Multivariate Methods 姓名 : 李政軒.
Review of statistical modeling and probability theory Alan Moses ML4bio.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
Pattern recognition – basic concepts. Sample input attribute, attribute, feature, input variable, independent variable (atribut, rys, příznak, vstupní.
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
Predictive Automatic Relevance Determination by Expectation Propagation Y. Qi T.P. Minka R.W. Picard Z. Ghahramani.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Computer vision: models, learning and inference
Latent Variables, Mixture Models and EM
Claudio Lottaz and Rainer Spang
CSCI 5822 Probabilistic Models of Human and Machine Learning
Elizabeth Garrett Giovanni Parmigiani
Hidden Markov chain models (state space model)
Computational Diagnostics
Rainer Spang, Max Planck Institute for Molecular Genetics, Berlin
Mathematical Foundations of BME Reza Shadmehr
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Biointelligence Laboratory, Seoul National University
OVERVIEW OF LINEAR MODELS
Robust Full Bayesian Learning for Neural Networks
Parametric Methods Berlin Chen, 2005 References:
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Claudio Lottaz and Rainer Spang
Presentation transcript:

Computational Diagnostics based on Large Scale Gene Expression Profiles using MCMC Rainer Spang, Max Planck Institute for Molecular Genetics, Berlin Harry Zuzan, Carrie Blanchette, Erich Huang, Holly Dressman, Jeff Marks, Joe Nevins, Mike West Duke Medical Center & Duke University

Estrogen Receptor Status 7000 genes 49 breast tumors 25 ER+ 24 ER-

Tumor – Chip - 7000 Numbers

Given Wanted 89% The probability that the tumor is ER+ 7000 Numbers

7000 Numbers Are More Numbers Than We Need Predict ER status based on the expression levels of super-genes

Singular Value Decomposition Singular values Loadings Data Expression levels of super genes, orthogonal matrix

Probit Model Class of tumor i Distribution Function of a Standard Normal Regression weight for super gene i Expression Level of super gene i

Overfitting Using only a small number of super genes is not robust at all When using many (all) supergenes, the linear model can be easily saturated, i.e. we have several models that fit perfectly well Consequence: For a new patient we find among these models some that support that she is ER+ and others that predict she is ER-

Given the Few Profiles With Known Diagnosis: The uncertainty on the right model is high The variance of the model-weights is large The likelihood landscape is flat We need additional model assumptions to solve the problem

Informative Priors Likelihood Prior Posterior

If the Prior Is Chosen Badly: We can not reproduce the diagnosis of the training profiles any more We still can not identify the model The diagnosis is driven mostly by the additional assumptions and not by the data

The Prior Needs to Be designed in 49 Dimensions Shape? Center? Orientation? Not to narrow ... not to wide

Shape multidimensional normal for simplicity

Assumptions on the model correspond to assumptions on the diagnosis Center Assumptions on the model correspond to assumptions on the diagnosis

Orientation orthogonal super-genes !

Not to Narrow ... Not to Wide Auto adjusting model Scales are hyper parameters with their own priors

Prior given the hyper parameter Rescaling by singular values Hyper parameter Independent super genes Unbiased prior

A prior for the hyper parameters Conjugate prior Flexibility for Symmetric U-Shaped prior for k=2 or k=3

Latent Variable Albert & Chip 1993

MCMC - Gibbs Sampler - Sequential updates of conditional distributions All conditional posteriors can be calculated analytically West 2001, Albert & Chip 1993

What are the additional assumptions that came in by the prior? The model can not be dominated by only a few super-genes ( genes! ) The diagnosis is done based on global changes in the expression profiles influenced by many genes The assumptions are neutral with respect to the individual diagnosis

Which Genes Have Driven the Prediction ? Weight nuclear factor 3 alpha 0.853 cysteine rich heart protein 0.842 estrogen receptor 0.840 intestinal trefoil factor x box binding protein 1 0.835 gata 3 0.818 ps 2 liv1 0.812 ... many many more ... ...

Thank you!