Parametric tests (independent t- test and paired t-test & ANOVA) Dr. Omar Al Jadaan.

Slides:



Advertisements
Similar presentations
2  How to compare the difference on >2 groups on one or more variables  If it is only one variable, we could compare three groups with multiple ttests:
Advertisements

Significance Testing Chapter 13 Victor Katch Kinesiology.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
PSY 307 – Statistics for the Behavioral Sciences
Statistics for Managers Using Microsoft® Excel 5th Edition
Biol 500: basic statistics
UNDERSTANDING RESEARCH RESULTS: STATISTICAL INFERENCE © 2012 The McGraw-Hill Companies, Inc.
Today Concepts underlying inferential statistics
Hypothesis Testing Using The One-Sample t-Test
Richard M. Jacobs, OSA, Ph.D.
Inferential Statistics
1 Nominal Data Greg C Elvers. 2 Parametric Statistics The inferential statistics that we have discussed, such as t and ANOVA, are parametric statistics.
Statistical Analysis. Purpose of Statistical Analysis Determines whether the results found in an experiment are meaningful. Answers the question: –Does.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Inferential Statistics
Chapter 12: Analysis of Variance
AM Recitation 2/10/11.
F-Test ( ANOVA ) & Two-Way ANOVA
Inferential Statistics: SPSS
Overview of Statistical Hypothesis Testing: The z-Test
Statistical Analysis Statistical Analysis
Section 10.1 ~ t Distribution for Inferences about a Mean Introduction to Probability and Statistics Ms. Young.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
The paired sample experiment The paired t test. Frequently one is interested in comparing the effects of two treatments (drugs, etc…) on a response variable.
The Hypothesis of Difference Chapter 10. Sampling Distribution of Differences Use a Sampling Distribution of Differences when we want to examine a hypothesis.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 22 Using Inferential Statistics to Test Hypotheses.
Chapter 9 Hypothesis Testing and Estimation for Two Population Parameters.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
One-sample In the previous cases we had one sample and were comparing its mean to a hypothesized population mean However in many situations we will use.
Hypothesis Testing Using the Two-Sample t-Test
Chapter 9: Non-parametric Tests n Parametric vs Non-parametric n Chi-Square –1 way –2 way.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
ANOVA (Analysis of Variance) by Aziza Munir
Copyright © Cengage Learning. All rights reserved. 10 Inferences Involving Two Populations.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Two-Sample Tests and One-Way ANOVA Business Statistics, A First.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 10-1 Chapter 10 Analysis of Variance Statistics for Managers Using Microsoft.
Chapter 12 A Primer for Inferential Statistics What Does Statistically Significant Mean? It’s the probability that an observed difference or association.
INTRODUCTION TO ANALYSIS OF VARIANCE (ANOVA). COURSE CONTENT WHAT IS ANOVA DIFFERENT TYPES OF ANOVA ANOVA THEORY WORKED EXAMPLE IN EXCEL –GENERATING THE.
Essential Question:  How do scientists use statistical analyses to draw meaningful conclusions from experimental results?
Chapter 9 Three Tests of Significance Winston Jackson and Norine Verberg Methods: Doing Social Research, 4e.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics ANalysis Of VAriance: ANOVA.
© Copyright McGraw-Hill 2000
Chapter Seventeen. Figure 17.1 Relationship of Hypothesis Testing Related to Differences to the Previous Chapter and the Marketing Research Process Focus.
Chapter Twelve The Two-Sample t-Test. Copyright © Houghton Mifflin Company. All rights reserved.Chapter is the mean of the first sample is the.
NON-PARAMETRIC STATISTICS
DTC Quantitative Methods Bivariate Analysis: t-tests and Analysis of Variance (ANOVA) Thursday 14 th February 2013.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved Lecture Slides Elementary Statistics Eleventh Edition and the Triola.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Chapter 8 Analysis of METOC Variability. Contents 8.1. One-factor Analysis of Variance (ANOVA) 8.2. Partitioning of METOC Variability 8.3. Mathematical.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 4 Investigating the Difference in Scores.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Nonparametric statistics. Four levels of measurement Nominal Ordinal Interval Ratio  Nominal: the lowest level  Ordinal  Interval  Ratio: the highest.
 List the characteristics of the F distribution.  Conduct a test of hypothesis to determine whether the variances of two populations are equal.  Discuss.
Chapter 10: The t Test For Two Independent Samples.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Two-Sample Hypothesis Testing
Lecture Slides Elementary Statistics Twelfth Edition
Lecture Slides Elementary Statistics Twelfth Edition
Chi Square (2) Dr. Richard Jackson
What are their purposes? What kinds?
Lecture Slides Elementary Statistics Twelfth Edition
Presentation transcript:

Parametric tests (independent t- test and paired t-test & ANOVA) Dr. Omar Al Jadaan

Student's t-test (t-test) The Student's t-test (or simply t-test) was developed by William Gosset - "Student" in 1908William Gosset - "Student" The t-test is used to compare two groups comes in at least 3 flavours: – Paired t-test: used when each data point in one group corresponds to a matching data point in the other group. – Unpaired t-test: used whether or not the groups contain matching data points: a)Two-sample assuming equal variances b)Two-sample assuming unequal variances

The t-test is a parametric test which assumes that the data analyzed:parametric test – Be continuous, interval data comprising a whole population or sampled randomly from a population.continuous, interval data – Have a normal distribution.normal distribution – If n 30).variances – Sample size should not differ hugely between the groups.

Paired t-test: The paired t-test is used to investigate the relationship between two groups where there is a meaningful one-to-one correspondence between the data points in one group and those in the other, e.g. a variable measured at the same time points under experimental and control conditions. It is NOT sufficient that the two groups simply have the same number of data points!

The advantage of the paired t-test is that the formula procedure involved is fairly simple: Start with the hypothesis (H 0 ) that the mean of each group is equal (H A : the means are not equal). How do we test this? By considering the variance (standard deviation) of each group. Set a value for a (significance level, e.g. 0.05). Calculate the difference for each pair (i.e. the variable measured at the same time point under experimental and controlled conditions). Plot a histogram of the differences between data pairs to confirm that they are normally distributed - if not, STOP! Calculate the mean of all the differences between pairs (d av ) and the standard deviation of the differences (SD) The value of t can then be calculated from the following formula:

where: d av is the mean difference, i.e. the sum of the differences of all the data points (set 1 point 1 - set 2 point 2,...) divided by the number of pairs SD is the standard deviation of the differences between all the pairs N is the number of pairs. N.B. The sign of t (+/-) does not matter, assume that t is positive. 7.The significance value attached to the resulting value of t can be looked up in a table of the t distribution (or obtained from appropriate software). To do this, you need to know the "degrees of freedom" (df) for the test. The degrees of freedom take account of the number of independent observations used in the calculation of the test statistic and are needed to find the true value in a probability table. For a paired t- test:table of the t distribution df = n-1 (number of pairs - 1)

To look up t, you also need to determine whether you are performing a one-tailed or two-tailed test.

Whether you use a one- or two-tailed test depends on your testing hypothesis: One-tailed test: Used where there is some basis (e.g. previous experimental observation) to predict the direction of the difference, e.g. expectation of a significant difference between the groups. Two-tailed test: Used where there is no basis to assume that there may be a significant difference between the groups - this is the test most frequently used.

If the calculated value of t is greater than the critical value, H 0 is rejected, i.e. there is evidence of a statistically significant difference between the groups. If the calculated value of t is less than the critical value, H 0 is accepted, i.e. there is no evidence of a statistically significant difference between the two groups.

Unpaired t-test: The unpaired t-test does not require that the two groups be paired in any way, or even of equal sizes. A typical example might be comparing a variable in two experimental groups of patients, one treated with drug A and one treated with drug B. Such situations are common in medicine where an accepted treatment already exists and it would not be ethical to withhold this from a control group. Here, we wish to know if the differences between the groups are "real" (statistically significant) or could have arisen by chance. The calculations involved in an unpaired t-test are slightly more complicated than for the paired test. Note that the unpaired t-test is equivalent to one-way ANOVA used to test for a difference in means between two groups. To perform an unpaired t-test:

1.Plot a histogram of the datasets to confirm that they are normally distributed - if not, STOP and use a non- parametric method! 2.Start with the hypothesis (H 0 ) "There is no difference between the populations of measurements from which samples have been drawn". (H A : There is a difference). 3.Set a value for a (significance level, e.g. 0.05). 4.Check the variance of each group. If the variances are very different, you can already reject the hypothesis that the samples are from a common population, even though their means might be similar. However, it is still possible to go on and perform a t-test.variance 5.Calculate t using the formula:

where: = means of groups A and B, respectively, and 6.Look up the value of t for the correct number of degrees of freedom and for a one- or two-tailed test (above). For an unpaired t-test: df = (n A + n B ) – 2 where n A/B = the number of values in the two groups being compared. Remember that the sign of t (+/-) does not matter - assume that t is positive.

7.If the calculated value of t is greater than the critical value, H 0 is rejected, i.e. there is evidence of a statistically significant difference between the groups. If the calculated value of t is less than the critical value, H 0 is accepted, i.e. there is no evidence of a statistically significant difference between the groups.

Example Consider the data from the following experiment. A total of 12 readings were taken, 6 under control and 6 under experimental conditions Under Control Under Experimental Mean: VARP SD

Are the data points for the control and experimental groups paired? - Nope, they are just replicate observations. Can we perform an unpaired t-test? Are the data normally distributed? Yes, approximately:

H 0 : "There is no difference between the populations of measurements from which samples have been drawn" (H A : There is a difference). Set value of a = 0.05 Are the variances of the two groups similar? - Yes (1.68 / 2.46) - less than 2-fold difference. Since all the requirements for a t-test have been met, we can proceed:

t = 1.08 Is this a one-tailed or a two-tailed test? - Two tailed, since we have no firm basis to assume that there is a difference between the groups. How many degrees of freedom are there? df = (N A -1)+(N B -1) = 10

From the table of values for t, we can see that for a two-tailed test with df=10 and a=0.05, t would have to be or greater for >5% (0.05) of pairs of samples to differ by the observed amount:table of values for t

If the calculated value of t is GREATER THAN the critical value of t (from the table), REJECT H 0. If the calculated value of t is LESS THAN the critical value of t (from the table), ACCEPT H 0. Since t calc = 1.08 and t crit = 2.228, the null hypothesis is accepted. CONCLUSION: There is no statistically significant difference (at the 95% confidence level) between the Experimental and the Control group

ANalysis Of VAriance: ANOVA ANOVA is a parametric test which assumes that the data analyzed:parametric test Be continuous, interval data comprising a whole population or sampled randomly from a population.continuous, interval data Has a normal distribution. Moderate departure from the normal distribution does not unduly disturb the outcome of ANOVA, especially as sample sizes increase. Highly skewed datasets result in inaccurate conclusions.normal distribution The groups are independent of each other. The variances in the two groups should be similar.variances For two-way ANOVA, the sample size the groups is equal (for one-way ANOVA, sample sizes need not be equal, but should not differ hugely between the groups).

One-way (or one-factor) ANOVA: Tests the hypothesis that means from two or more samples are equal (drawn from populations with the same mean). Student's t-test is actually a particular application of one-way ANOVA (two groups compared) and results in the same conclusions.

Two-way (or two-factor) ANOVA: Simultaneously tests the hypothesis that the means of two variables ("factors") from two or more groups are equal (drawn from populations with the same mean), e.g. the difference between a control and an experimental variable, or whether there is a difference between alcohol consumption and liver disease in several different countries. Does not include more than one sampling per group. This test allows comments to be made about the interaction between factors as well as between groups.

The F ratio The F (Fisher) ratio compares the variance within sample groups ("inherent variance") with the variance between groups ("treatment effect") and is the basis for ANOVA:variance F = variance between groups / variance within sample groups

Example One-way ("Single Factor") ANOVA: "Pain Score" for 3 Analgesics: Aspirin: Paracetemol (Acetaminophen) : Ibuprophen: Control (no drug):

The null hypothesis (H 0 ) is that there is no difference between the four groups being compared. In this example, with a significance level of 95% (a = 0.05), since the calculated value of F (3.23) is greater than F crit (3.01), we reject the null hypothesis that the three drugs perform equally.

A post-hoc comparison or individual pair wise comparisons would have to be be performed to determine which pair or pairs of means caused rejection of the null hypothesis.

Example Two-way ANOVA ("Two-factor without replication"): Apple codling moth (Cydia pomonella) caught in pheromone traps: Bait 1:Bait 2: Orcha rd 1: Orcha rd 2:

As always, the null hypothesis (H 0 ) is that there is no difference between the groups being compared. In this example, with a significance level of 95% (a = 0.05), the calculated value of F (10.57) for the table rows (Orchard 1 vs. Orchard 2) is greater than F crit (2.82), so the hypothesis that there is no difference between the orchards is rejected. However, the calculated value of F (0.48) for the table columns (Bait 1 vs. Bait 2) is less than F crit (2.82), so the hypothesis that there is no difference between the pheromone baits is accepted.