CHI SQUARE DISTRIBUTION
The Chi-Square ( 2 ) Distribution The chi-square distribution is the probability distribution of the sum of several independent, squared standard normal random variables.
The Chi-Square ( 2 ) Distribution The chi-square random variable cannot be negative, so it is bound by zero on the left. The chi-square distribution is skewed to the right. The chi-square distribution approaches a normal as the degrees of freedom increase.
Area in Right Tail Area in Left Tail df Values and Probabilities of Chi-Square Distributions
Test of Goodness of Fit It enable us to determine how good a fit is between observed and expected frequencies. Where the former is the outcome of samples, the later are determined from the hypothesized population from which the sample is taken. Conventionally the Null hypothesis is formed that says there is no difference between the obtained and expected frequencies, i.e there is a good between observed and expected.
Test of Independence Here we try to test whether two different types of attribute of same population are statistically independent or not. Conventionally the Null hypothesis is formed that says there is no difference between the obtained and expected frequencies. In other words attributes under study are independent of each other.
The Chi-Square Distribution The chi-square ² distribution depends on the number of degrees of freedom A chi-square point ² α is the point under a chi-square distribution that gives right-hand tail area
14-9 A Chi-Square Test for Goodness of Fit Steps in a chi-square analysis: Formulate null and alternative hypotheses Compute frequencies of occurrence that would be expected if the null hypothesis were true - expected cell counts Note actual, observed cell counts Use differences between expected and actual cell counts to find chi-square statistic: Compare chi-statistic with critical values from the chi-square distribution (with k-1 degrees of freedom) to test the null hypothesis Steps in a chi-square analysis: Formulate null and alternative hypotheses Compute frequencies of occurrence that would be expected if the null hypothesis were true - expected cell counts Note actual, observed cell counts Use differences between expected and actual cell counts to find chi-square statistic: Compare chi-statistic with critical values from the chi-square distribution (with k-1 degrees of freedom) to test the null hypothesis
The null and alternative hypotheses: H 0 : The probabilities of occurrence of events E 1, E 2...,E k are given by p 1,p 2,...,p k H 1 : The probabilities of the k events are not as specified in the null hypothesis The null and alternative hypotheses: H 0 : The probabilities of occurrence of events E 1, E 2...,E k are given by p 1,p 2,...,p k H 1 : The probabilities of the k events are not as specified in the null hypothesis Assuming equal probabilities, p 1 = p 2 = p 3 = p 4 =0.25 and n=80 PreferenceTanBrownMaroonBlackTotal Observed Expected(np) (O-E) Goodness-of-Fit Test
z f ( z ) Partitioning the Standard Normal Distribution Use the table of the standard normal distribution to determine an appropriate partition of the standard normal distribution which gives ranges with approximately equal percentages. p(z<-1) = p(-1<z<-0.44)= p(-0.44<z<0)= p(0<z<0.44)= p(0.44<z<14)= p(z>1) = Given z boundaries, x boundaries can be determined from the inverse standard normal transformation: x = + z = z. 3. Compare with the critical value of the 2 distribution with k-3 degrees of freedom. Goodness-of-Fit for the Normal Distribution
iO i E i O i - E i (O i - E i ) 2 (O i - E i ) 2 / E i 2 : iO i E i O i - E i (O i - E i ) 2 (O i - E i ) 2 / E i 2 : 2 (0.10,k-3) = > H 0 is not rejected at the 0.10 level Solution
Test for Independence
Null and alternative hypotheses: H 0 : The two classification variables are independent of each other H 1 : The two classification variables are not independent Chi-square test statistic for independence: Degrees of freedom: df=(r-1)(c-1) Expected cell count: A and B are independent if:P(A B) = P(A) P(B). If the first and second classification categories are independent:E ij = (R i )(C j )/n A and B are independent if:P(A B) = P(A) P(B). If the first and second classification categories are independent:E ij = (R i )(C j )/n Contingency Table Analysis: A Chi-Square Test for Independence
ijOEO-E(O-E) 2 (O-E) 2 /E 2 : 2 (0.01,(2-1)(2-1)) = H 0 is rejected at the 0.01 level and it is concluded that the two variables are not independent. Contingency Table Analysis:
F - Distribution When independent samples of size n1 and n2 are drawn from two normal population, the ratio Follows F- distribution with n1-1 and n2-1 df. Where s1 and s2 are sample variance and
Test of Independence Here we try to test whether two different types of attribute of same population are statistically independent or not. Conventionally the Null hypothesis is formed that says there is no difference between the obtained and expected frequencies. In other words attributes under study are independent of each other.
Test of Independence Here we try to test whether two different types of attribute of same population are statistically independent or not. Conventionally the Null hypothesis is formed that says there is no difference between the obtained and expected frequencies. In other words attributes under study are independent of each other.
Test of Independence Here we try to test whether two different types of attribute of same population are statistically independent or not. Conventionally the Null hypothesis is formed that says there is no difference between the obtained and expected frequencies. In other words attributes under study are independent of each other.