Differential Expressions Classical Methods Lecture Topic 7.

Slides:



Advertisements
Similar presentations
Introductory Mathematics & Statistics for Business
Advertisements

Tests of Hypotheses Based on a Single Sample
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Statistical Modeling and Data Analysis Given a data set, first question a statistician ask is, “What is the statistical model to this data?” We then characterize.
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
Inference Sampling distributions Hypothesis testing.
From the homework: Distribution of DNA fragments generated by Micrococcal nuclease digestion mean(nucs) = bp median(nucs) = 110 bp sd(nucs+ = 17.3.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved.
Chapter 10: Hypothesis Testing
Significance Testing Chapter 13 Victor Katch Kinesiology.
Detecting Differentially Expressed Genes Pengyu Hong 09/13/2005.
QM Spring 2002 Business Statistics Introduction to Inference: Hypothesis Testing.
10 Hypothesis Testing. 10 Hypothesis Testing Statistical hypothesis testing The expression level of a gene in a given condition is measured several.
An Inference Procedure
Differentially expressed genes
Analysis of Differential Expression T-test ANOVA Non-parametric methods Correlation Regression.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
Chapter 9 Hypothesis Testing.
Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Inference about Population Parameters: Hypothesis Testing
Different Expression Multiple Hypothesis Testing STAT115 Spring 2012.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
False Discovery Rate (FDR) = proportion of false positive results out of all positive results (positive result = statistically significant result) Ladislav.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
Hypothesis Testing:.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Probability Distributions and Test of Hypothesis Ka-Lok Ng Dept. of Bioinformatics Asia University.
Multiple testing in high- throughput biology Petter Mostad.
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
Tuesday, September 10, 2013 Introduction to hypothesis testing.
Fundamentals of Hypothesis Testing: One-Sample Tests
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Intermediate Applied Statistics STAT 460
More About Significance Tests
Essential Statistics in Biology: Getting the Numbers Right
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Chapter 8 Introduction to Hypothesis Testing
Introduction to Hypothesis Testing: One Population Value Chapter 8 Handout.
STA Statistical Inference
Exam 1 Median: 74 Quartiles: 68, 84 Interquartile range: 16 Mean: 74.9 Standard deviation: 12.5 z = -1: 62.4 z = -1: 87.4 z = -1z = +1 Worst Question:
Differential Expression II Adding power by modeling all the genes Oct 06.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
10.2 Tests of Significance Use confidence intervals when the goal is to estimate the population parameter If the goal is to.
Multiple Testing in Microarray Data Analysis Mi-Ok Kim.
CHAPTER 17: Tests of Significance: The Basics
Bioinformatics Expression profiling and functional genomics Part II: Differential expression Ad 27/11/2006.
Back to basics – Probability, Conditional Probability and Independence Probability of an outcome in an experiment is the proportion of times that.
1 Chapter 8 Hypothesis Testing 8.2 Basics of Hypothesis Testing 8.3 Testing about a Proportion p 8.4 Testing about a Mean µ (σ known) 8.5 Testing about.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Copyright ©2013 Pearson Education, Inc. publishing as Prentice Hall 9-1 σ σ.
CHAPTER 15: Tests of Significance The Basics ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
Chapter 8 Parameter Estimates and Hypothesis Testing.
Fall 2002Biostat Statistical Inference - Confidence Intervals General (1 -  ) Confidence Intervals: a random interval that will include a fixed.
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 13: One-way ANOVA Marshall University Genomics Core.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
1 Definitions In statistics, a hypothesis is a claim or statement about a property of a population. A hypothesis test is a standard procedure for testing.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 6 –Multiple hypothesis testing Marshall University Genomics.
Chapter 12 Tests of Hypotheses Means 12.1 Tests of Hypotheses 12.2 Significance of Tests 12.3 Tests concerning Means 12.4 Tests concerning Means(unknown.
BIOL 582 Lecture Set 2 Inferential Statistics, Hypotheses, and Resampling.
Hypothesis Tests for 1-Proportion Presentation 9.
Chapter 9 Introduction to the t Statistic
Estimating the False Discovery Rate in Genome-wide Studies BMI/CS 576 Colin Dewey Fall 2008.
Introduction to Hypothesis Testing: The Binomial Test
Introduction to Hypothesis Testing: The Binomial Test
Chapter 9 Hypothesis Testing.
Presentation transcript:

Differential Expressions Classical Methods Lecture Topic 7

What’s your question? What are the genes that are different for the healthy versus diseased cells? –Gene discovery, differential expression Is a specified group of genes all up-regulated in a specified condition? –Gene set differential expression Can I use the expression profile of cancer patients to predict chemotherapy outcome? –Class prediction, classification Are there tumour sub-types not previously identified? Do my genes group into previously undiscovered pathways? –Class discovery, clustering This lecture covers first question - differential expression.

Some terms The following terms are used exchangeably: Differentially versus not differentially expressed Active versus Inactive genes Affected versus not affected genes We will first deal with the simplest case, comparing 2 conditions (say treatment and control). We will also mainly be interested in hypothesis testing (HT) aspect of Inference

Inference Lets first differentiate between the different schools of thought: Classical Inference Bayesian Inference Empirical Bayes’ methods

The MAIN Difference The Classical School believes that PARAMETERS are unknown but FIXED quantities and the data are random The Bayesians believe after observing, data can be considered fixed, whereas our knowledge about the parameter is truly random. This leads to two vastly different APPROACHES when it comes to hypothesis testing, though both approaches are inherently interested in the parameter.

Consider the following situation  g = population mean difference in gene expression from treatment to control for gene g. Both approaches are interested in this. Particularly, both approaches are interested in seeing if the parameter is 0 or not, i.e. whether or not the genes are differentially expressed. This corresponds to information about

The difference Classical Statisticians tend to be drastic about the choice of I. It is either 0 or 1. Bayesians on the other hand tend to describe this value with a probability, just as if it were another parameter. The classical approach to (HT) in micro-array is to assume the negative hypothesis (null or H0) and try to find the absurdity probability of the null hypothesis. The Bayesian approach assumes that for each gene g, there is the unobservable function, vg, that defines the genes activation status. The idea here is to update the knowledge of v g in terms of a posterior probability. This is the probability that v g =0.

The Classical Procedure Here we have some claim or belief or knowledge or guess about a population parameter which we want to prove (or disprove) using our data for evidence. Hypothesis is always written as a pair. Research: what we are trying to prove (H a ) Null: that nullifies / negates our research.(H 0 )

The hypotheses H 0g : gene g is NOT differentially expressed H 1g : gene g IS differentially expressed. We have n pairs of mutually exclusive hypothesis. The role of the two hypothesis is NOT symmetric. Classical Statistics assumes the null by default UNLESS enough contrary evidence is found. Hypothesis is a systematic procedure to summarize evidence in the data, in order to decide between the pair of hypothesis.

The Test Statistic The test statistic is a summary of the data used to evaluate the validity of the hypothesis. The summary chosen depends upon the question and circumstances. Typically the Test Statistic is chosen, so that when H0 is true the distribution of the test statistic has a certain known distribution.

The logic of hypothesis testing To actually test the hypothesis, what we try to do is to disprove or reject the null hypothesis. If we can reject the null, by default our Ha (which is our research) is true. How do we do this? We take a sample and look at the test-statistic. Then we try to quantify that, if the null was true, would our observed statistic be a likely value from the known distribution of test statistics. If our observed value is not a likely value, we reject the null. How likely or unlikely a value is, is determined by the sampling distribution of that statistic.

Error Rates Since we take our decisions about the parameter based on sample values we are likely to commit some errors. Type I error: Rejecting Ho when it is true Type II error: Failing to reject Ho when Ha is true. In any given situation we want to minimize these errors. P(Type I error) = , Also called size, level of significance. P(Type II error) =  Power = 1-  HERE we reject Ho when the claim is true. We want power to be LARGE

Terms in error control In hypothesis we have two main types of error: Null hypothesis is wrongly rejected (FALSE POSITIVE) Null hypothesis is wrongly accepted (failed to reject) (FALSE NEGATIVE). Trade off is one cannot control both probabilities of both errors together. Size = Probability of a False Positive Power= 1 – Probability of a False Negative

Decision Rule A decision rule is a method than translates the vector of observed values into a binary decision or REJECT or FAIL TO REJECT. The decision rule is made such that the error rate of Type I error is controlled. P-value: the absurdity value of the null hypothesis, or the probability of observing a number more extreme than the observed value, assuming that the null is true.

Classical Test for differential Expression concerning two populations The two sample t test –Version 1: The Pooled t test –Version 2: The Satterwaite Welch’s t test –Version 3: Bootstrapped t tests –Version 4: Permutation t test –Wilcoxon Rank-Sum test (nonparametric alternative) –Likelihood Ratio Test

The Test Statistic: In general Notation: Let x gi1,…,x gik represent the observations (intensity/ log intensity/ normalized intensity etc) for gene g, for the ith condition with, i=1,2. Let the means of the two conditions, the standard deviations and the variance be given by

t-test The test statistic is given by: This follows a t distribution with (k1+k2-2) df

The Satterwhite-Welch’s (SW) Test Statistic

The distribution of the Test Statistic The pooled t assumes that the two variances are equal. The SW test does not. Both assume underlying normality for the observations x gi1 …x giki.

Bootstrapping and Permutation t Tests Before we do the tests lets talk about the procedures. The tests use the principles of Bootstrapping or Permutation and apply it in the case of differential expression for two conditions.

Bootstrapping The term “bootstrap” came from the legendary Baron Munchausen, who pulled himself out of a man-hole literally by grabbing his own bootstraps. Efron (1979) introduced the idea into Statistics, the idea is to “get yourself out of a hole” by re-using the data many times. It is essentially a re-sampling technique that simulates alternative values of the statistic under the null (in our case the t statistic), in order to calculate an empirical p-value. The idea is to create a large number of values from the statistic to get an idea of the distribution of the test statistic UNDER the null.

Bootstrapping contd… First I will use a small example for illustrating the procedure and then I will introduce notation. Lets say we have data on gene “g” for condition 1 and condition 2 as follows: Cond Cond

Bootstrapping Calculate the observed t value for your data First you pool the data: Then samples of size 4 with replacement from the pooled sample, one for sample 1 and the other for sample 2. Calculate the t statistic Continue this for B samples. Find what proportion of observations are BIGGER than your observed value. This is your “estimated p-value” using bootstrap.

Example s1s2combs1bs2b

Permutation Test Similar Idea to Bootstrapping. Here instead of resampling WITH replacement we “shuffle” the data to look at all possible permutations. Hence, we pool the data, then sample without replacement from the pooled data. Same Idea, create a gamut of t statistics to get the empirical distribution of the null.

Permutation Test Contd We should ONLY use permutation tests when the number of replicates are >10, else the number of possible resamples is quite small. Example when n1=4, n2=4. The possible permutation is 8!/4!4!=70 possibilities, which is too small to estimate empirical p-values. Also, we need a firm assumption that the ONLY difference between the two samples is due to location and they have the same shapes.

Wilcoxon Rank Sum Test The non-parametric alternative to the two-sample t. Doesn’t assume underlying normality Pool the data and rank all the observations in the pooled data. Then sum the ranks of ONE of the samples. Idea is, if there is a location difference among the samples, the sum of ranks for one sample will be either bigger or smaller than the other. Wilcoxon Rank Sum tables are generally available. Not very good for VERY few replicates. At least 4, preferably 6.

Large Sample Approx. Here n1 is size of the sample we summed up.

Likelihood Ratio Test The idea of Likelihood Ratio Test is simple and intuitive. One basically looks at the ratio of the likelihood that the gene is differentially expressed to that of the likelihood it isn’t. The method was used by Idekar et al (2000). Model: x g1k =  g1 +  g1  g1k +  g1k x g2k =  g2 +  g2  g2k +  g2k Allows both additive and multiplicative error. Assumption,  gik,  g2k = BVN( 0, 0,  2  g1,  2  g2,   )  g1k,  g2k = BVN( 0, 0,  2  g1,  2  g2,   ) Allows correlations.

The LR Statistic The parameters  = (  2  g1,  2  g1,    2  g1,  2  g1,    and  =(  g1,  g2 ) are estimated via ML techniques optimizing: The LR Statistic is

The Issues of MULTIPLE Testing In Microarray setting there are often thousands of tests to be considered simultaneously. Obviously we have a pretty high chance of having false positives when you do thousands of tests at the same time. In these kinds of situations, the question of which error to control becomes an issue. We would like to control both Type I and Type II errors, but that’s not possible –If you make Type I error smaller, Type II goes up and vice versa. The general practice is to fix Type I error at a low level (.05,.01,.001) and minimize the Type II error for that level.

Numbers of Correct and Incorrect Decisions Declared Inactive Declared Active Total Truly Inactive UV =F P n0n0 Truly Active T=F N S=T P n-n 0 Totaln-RRn

What Error Rates do we need to control? Most traditional methods focus on controlling the expected fraction of false positives out of the total number of TRUE null hypotheses. We want the expected fraction of false positives to be small. We want to control FPR = E ( F P / n 0 ) However since we do not know n 0 we really cannot control this rate. So we look at several measures that are related to this rate.

Per-comparison Error Rate (PCER): PCER = E(V)/n Controlling the expected number of False positives as a fraction of the total number of hypothesis tested. Here a level control achieved by performing each test in the family at a level alpha error arising from multiplicity ignored false claims of significance are made at a higher rate than

Per-family Error Rate (PFER): PFER = E(V) The other end of the spectrum: controls the expected number of False positive per each hypothesis tested. This is the expected number of False Positives in a study. procedures that control this error rate are very strict rarely used in microarray literature

Family-wise Error Rate (FWER): FWER = P(V ≥ 1) Controls the probability of having a single Type I error in the study. Most commonly used, common examples are: Bonferroni, Tukey, Dunnett procedures based on this error rate are very conservative generalized version of this error rate: k-FWER

False Discovery Rate (FDR): FDR = E(V/R | R ≥ 1) Controls the expected number of False positives OUT OF THE ONES DECLARED POSITIVE. FDR controlling procedures are liberal widely used in microarray literature

The Methods Compared The hierarchy among techniques PCER ≥ FDR ≥ FWER ≥ PFER PCER is the least conservative and PFER is the most conservative. These days people use either FWER or FDR These are all SINGLE STEP methods. However, Step- wise methods are getting more popular.

Stepwise methods Essentially a more adaptive form of testing. We can have step-down methods or step-up methods. Step-down methods: rank observed p-values from the smallest to the highest, –p (1) ≤ … ≤p (m) –Start at the smallest p-value, p (1). Step-Up Method –Rank as before –start with the largest p-vlaue

Step-Down: Bonferroni Holm Method rank observed p-values from the smallest to the highest, p (1) ≤ … ≤p (m) and define the corresponding H 1 ….H m Start with the smallest one, reject H 1 if p (1) <  /m and continue to Step 2 using p (2) and so on… Else Fail to reject all H 1 ….H m. And reject H (j) if p (j) <  (m - j + 1) This is more powerful than single-step procedures

Step-up: Hochberg and Hommel rank observed p-values from the smallest to the highest, p (1) ≤ … ≤p (m) and define the corresponding H 1 ….H m Start with the LARGEST one, reject H m if p (m) <  and all the corresponding H 1 ….H m Otherwise Fail to reject and continue to Step 2 using p (m-1) and so on… And reject H (j) if this procedure rejects H(1),...,H(k) if k =Max{j| p(j) ≤ j  m} exists. This is more powerful than single-step procedures and the Step-Down procedure and controls FDR

Error Rates: used in microarrays FWER Bonferroni: Classify all genes as active if their p-value is less than  n Step-up FWER(Hochberg): Let k be the largest g =1,…n, for which p (g) <=  (n-g+1). Then reject all H (g) for g=1…k where H (g) is the null associated with the gth smallest p-value. Step-down FWER(Holmes): Let k be the smallest g =1,…n, for which p (g) >=  (n-g+1). Then fail to reject all H (g) for g=1…k where H (g) is the null associated with the gth largest p-value FDR(Benjamini and Hochberg): Let k be the largest g =1,…n, for which p (g) <=  g/np 0 ). Then reject all H (g) for g=1…k where H (g) is the null associated with the gth smallest p-value. (here p0 is the true proportion of inactive genes which is unknown, and generally replaced by 1 a conservative value)