Download presentation
Presentation is loading. Please wait.
1
The Need For Resampling In Multiple testing
2
Correlation Structures Tukey’s T Method exploit the correlation structure between the test statistics, and have somewhat smaller critical value than the Bonferroni-style critical values. It is easier to obtain a statistically significant result when correlation structures are incorporated.
3
Correlation Structures The incorporation of correlation structure results in a smaller adjusted p-value than Bonferroni-style adjustment, again resulting in more powerful tests. The incorporation of correlation structures can be very important when the correlations are extremely large.
4
Correlation Structures Often, certain variables are recognized as duplicating information and are dropped, or perhaps the variables are combined into a single measure. In the case, the correlations among the resulting variables is less extreme.
5
Correlation Structures In cases of moderate correlation structures, the distribution between the Bonferroni adjustment and the exact adjustment can be very slight. Bonferroni inequality: Pr{∩ 1 r A i } ≧ 1-Σ 1 r Pr{A i c } A small value of ‘’Pr{A i c }’’corresponds to a small per-comparison error rate.
6
Correlation Structures The incorporating dependence structure becomes less important for smaller significant levels. If a Bonferroni-style correction is reasonable, then why bother with resampling?
7
Distributional Characteristics Other distributional characteristics, such as discreteness and skewness, can have dramatic effect, even for small p-value. The nonnormality is of equal or greater concern than correlation structure in multiple testing application.
8
The Need For Resampling In Multiple testing Distribution Of Extremal Statistics Under Nonnormality
9
Noreen’s analysis of tests for a single lognormal mean Y ij are observations. i=1,..,10, j=1,..,n All observations are independent and identically distributed as e z, where Z denotes a standard normal random variable. The hypotheses tested are H i : E(Y ij )=√e, with upper or lower-tailed alternatives. t=(y-√e)/(s/√n) _
10
Distributions of t-statistics For each graph there were 40000 t-statistics, all simulated using lognormal y ij. The solid lines (actual) show the distribution of t when sampling from lognormal population, and the dotted lines (nominal) show the distribution of t when sampling from normal population.
11
Distributions of t-statistics
12
The lower tail area of the actual distribution of t-statistic is larger than the corresponding tail of the approximating Student’s t- distribution, the lower-tailed test rejects H more often than it should. The upper tail area of the actual distribution is smaller than that of the approximating t- distribution, yielding fewer rejections than expected.
13
Distributions of t-statistics As can be expected with larger sample sizes, the approximations become better, and the actual proportion of rejections more closely approximates the nominal proportion.
14
Distributions of minimum and maximal t-statistics When one considers maximal and minimal t- statistics, the effect of the skewness is greatly amplified.
15
Distributions of minimum t-statistics
16
Distributions of minimum t-statistics (lower-tail) Because values in the extreme lower tails of the actual distributions are much more likely than under the corresponding t-distribution, the possibility of observing a significant result can be much larger than expected under the assumption of normal data. This cause false significances.
17
Distributions of minimum t-statistics (upper-tail) It is quit difficult to achieve a significant upper-tailed test, since the true distributions are so sharply curtailed in the upper tails. It has very lower power, and will likely fail to detect alternative hypotheses.
18
Distributions of maximum t-statistics
19
Distributions of minimum and maximal t-statistics We can expect that these results will become worse as the number of tests (k) increases.
20
The normal-based tests are much robust when testing contrasts involving two or more groups. T=(Y 1 -Y 2 )/s√(1/n 1 +1/n 2 ) _ Two-sample Tests _
21
There is an approximate cancellation skewness terms for the distribution of T, leaving the distribution roughly symmetric. We expected the normal-based procedures to perform better than in the one-sample case.
22
Two-sample Tests According to the rejection proportions, both procedures perform fairly well. Still, the bootstrap performs better than the normal approximation.
23
The Need For Resampling In Multiple testing The performance of Bootstrap Adjustments
24
Bootstrap Adjustments Use the adjusted p-values for the lower-tailed tests The pivotal statistics used to test the ten hypotheses are
25
Bootstrap Adjustments For Ten Independent Samples
26
Bootstrap Adjustments The adjustment algorithm in Algorithm 2.7 was placed within an ‘outer loop’, in which the data y ij were repeatedly generated iid from the standard lognormal distribution.
27
Bootstrap Adjustments We generate NSIM=4000 data sets, all under the complete null hypothesis. For each data set, we computed the bootstrap adjusted p-value using NBOOT 1000 bootstrap samples. The proportion of the NSIM samples having an adjusted p-value below α estimates the true FEW level of the method.
28
Rejection Proportions
29
The bootstrap adjustments The bootstrap adjustments are much better approximation. The bootstrap adjustments may have fewer excess Type I errors than the parametric Sidak adjustments. (lower-tail) The bootstrap adjustments may be more powerful than the parametric Sidak adjustments. (upper-tail)
30
Step-down Methods For Free Combination
31
Step-down methods Rather than adjust all p-values according to the min P j distribution, only adjust the minimum p-value using this distribution. Then adjust the remaining p-values according to smaller and smaller sets of p- value. It makes the adjusted p-value smaller, thereby improving the power of the single- step adjustment method.
32
Free combinations If, for every subcollection of j hypotheses {H i1,..,H ij }, the simultaneous truth of {H i1,..,H ij } and falsehood of the remaining hypotheses is plausible event, then the hypotheses satisfy the free combinations condition. In other words, each of the 2 k outcomes of the k-hypothesis problem is possible.
33
Holm’s method (Step-down methods)
34
Boferroni Step-down Adjusted p- values An consequence of the max adjustment is that the adjusted p-values have the same monotonicity as the original p-values.
35
Example Consider a multiple testing situation with k=5 where the ordered p-values p (i) are 0.009,0.011,0.012,0.134, and 0.512. Let H (1) be the hypothesis corresponding to the p-value 0.0009, H (2) be the hypothesis corresponding to 0.011, and so on. α=0.05
36
Example
37
Monotonicity enforcement In stages 2 and 3, the adjusted p-values were set equal to the first adjusted p- value,0.045. Without such monotonicity enforcement, the adjusted p-values p 2 and p 3 would be smaller than p 1. One might accept H (1) yet reject H (2) and H (3). It would run contrary to Holm’s algorithm.
38
Bonferroni Step-down Method Using the single-step method, the adjusted p-values are obtained by multiplying every raw p-value by five. Only H (1) test would be declared significant at the FEW=0.05. The step-down Bonferroni method is clearly superior to the single-step Bonferroni method. Slightly less conservative adjustments are possible by using the Sidak inequality, taking the adjustments to be 1-(1-p (j) ) (k-j+1) at step j.
39
The free step-down adjusted p-values (Resampling) The adjustments may be made less conservative by incorporating the precise dependence characteristics. Let the ordered p-values have indexes r 1,r 2,…,so that p (1) =p r1,p (2) =p r2,…,p (k) =p rk
40
The free step-down adjusted p-values (Resampling)
41
The adjustments are uniformly smaller than the single-step adjusted p-values, since the minima are taken over successively smaller sets.
42
Free Step-down Resampling Method
44
Example K=5 P-values are 0.009, 0.011, 0.012, 0.134, and 0.512. Suppose these correspond to the original hypotheses H 2,H 4,H 1,H 3, and H 5.
45
A Specific Step-down Illustration
47
Step-Down Methods For Restricted Combinations
48
When the hypotheses are restricted, then certain combinations of true hypotheses necessarily imply truth or falsehood of other hypotheses. In these cases, the adjustments may be made smaller than the free step-down adjusted p-values.
49
Step-Down Methods For Restricted Combinations The restricted step-down method starts with the ordered p-values,p (1) ≦ … ≦ p (k),p (j) =p rj. If H (j) is rejected, then H (1),…,H (j-1) must have been previously rejected. The multiplicity adjustment for the restricted step-down method at stage j considers only those hypotheses that possibly can be true, given that the previous j-1 hypotheses are all false.
50
Step-Down Methods For Restricted Combinations Define sets s j of hypotheses which include H (j) that can be true at stage j, given that all previous hypotheses are false. S={r 1,…,r k }={1,…,k}, define __
51
The Bonferroni adjustments Define |K|= the number of elements in the set K.
52
Step-Down Methods For Restricted Combinations(Bonferroni) The adjusted p-values can be no larger than the free Bonferroni adjustments, since M j ≦ k-j+1. In the case of free combinations, the truth of a collection of null hypotheses indexed by {r j,…,r k } cannot contradict the falsehood of all nulls indexed by {r 1,..,r j-1 }. In this case, S j ={{r j,…,r k }}, thus M j =k-j+1,and the restricted method reduces to the free method as a special case.
53
Step-Down Methods For Restricted Combinations(resampling)
54
At each step of (2.13),the probabilities are computed over subsets of the sets in (2.10). Thus, the restricted adjustments (2.13) can be no larger than the free adjustments.
55
Error Rate Control For Step-Down Method Error Rate Control Under H o k
56
The probability of rejecting at least one H 0i is no larger than α, no matter what subset of the K of hypotheses happens to be true. K o ={i 1,…,i j } denote the collection of hypotheses H0i which are true. Let x k α denote theα quantile of min P t | H o c
57
Error Rate Control Under H o k Define
58
Critical Value-Based Sequentially Rejective Algorithm
59
Error Rate Control Under H o k We have the following relationships: Where j ≦ k-|K 0 |+1 is defined by min P t =P (j) =P rj
60
Error Rate Control Under H o k
61
which demonstrates that the restricted step- down adjustments strongly control the FEW.
62
Error Rate Control Under H k Suppose that H k is true, then the distribution of Y is G. Suppose also that there exist random variables P i0, defined on the same probability space as the Pi, for which P i ≧ P i0 for all i.
63
Error Rate Control Under H k The error rate is controlled:
64
Error Rate Control Under H k Such P i0 frequently exist in parametric analyses; for example, the two-sample t- statistic for testing H 0 :μ 1 ≦ μ 2 may be written:
65
Error Rate Control Under H k The p-value for this test is p=Pr(T 2(n-1) ≧ t). Letting the p 0 be defined by p 0 =Pr(T 2(n-1) ≧ t 0 ), p 0 < p whenever μ 1 < μ 2.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.