Presentation is loading. Please wait.

Presentation is loading. Please wait.

Inferential Statistics Introduction. If both variables are categorical, build tables... Convention: Each value of the independent (causal) variable has.

Similar presentations


Presentation on theme: "Inferential Statistics Introduction. If both variables are categorical, build tables... Convention: Each value of the independent (causal) variable has."— Presentation transcript:

1 Inferential Statistics Introduction

2 If both variables are categorical, build tables... Convention: Each value of the independent (causal) variable has its own column One table: insert the observed frequencies (# cases that share corresponding values of the ind. and dep. variables) Another table: insert %. Compute separately for each value of the independent varaible (columns total 100%) Bivariate (two variables) Additional independent (control) variables: Construct a first-order partial table for each level of the new variable Multivariate (three or more variables)

3 r =.87 r 2 =.76 Hypothesis: POVERTY  CRIME For additional (control) variables, use partial correlation or multiple regression If both variables are continuous, use the r (correlation) and r 2 (regression) statistics Bivariate (two variables) Multivariate (three or more variables)

4 Inferential statistics Go beyond simple evaluations, such as comparing proportions or calculating a correlation (r) Inferential statistics allow us to legitimately “generalize” our findings – apply the results from a sample to a population Since we are projecting our findings to a population, we must draw samples. If our sample is a population, we cannot use these methods. We must use probability sampling (i.e., random sampling) We calculate a “test statistic”, such as an r, X 2, t, F, Beta, etc. If this statistic is sufficiently large we can say that there is a “statistically significant” relationship between variables.

5 How inferential statistics work Always based on a POPULATION from which a probability (preferably – random) sample is taken Excepting Chi-Square, which has its own “sampling” distribution, it is assumed that the dependent variable scores in the population are normally distributed around a mean or centerpoint This centerpoint is the score on the dependent variable that one would expect by randomly drawing a single case from the population To assess the effect of independent variables, the computer analyzes the dependent variable scores for a particular sample. For example, in the difference between means test, the centerpoint is the average difference between all possible sample means. (This is automatically computed.) If the difference between a sample mean and the centerpoint is large enough to wind up in the distribution’s “tail” (shaded area), the relationship is deemed statistically “significant”. Essentially the same logic applies to all inferential statistics.

6 Systematic and Error Effects Whenever we “observe” that one variable may be causing corresponding changes in another variable, we are really seeing the sum of two things: –A “systematic” effect, meaning that portion of the relationship that is produced by the action of the independent variable –An “error” or “chance” effect, meaning that portion that is produced by random factors Imagine that the entire surface area of this shape represents the observed relationship between two variables How does removing what is due to chance affect our estimate of the strength of the relationship between the variables? Imagine that our calculations determine that this portion of the relationship is due to chance

7 The null hypothesis holds that the relationship between variables (including any difference between groups) is caused by chance. Essentially, that our working hypothesis is incorrect. To “reject” the null hypothesis we must demonstrate that the association between variables is substantially larger than would be produced by chance alone. The first step is to compute a “test statistic” – an r, X 2, t, F, Beta, etc. Based on the size of the test statistic, the sample size, etc. the computer calculates the probability that the null hypothesis is true. If at least ONE asterisk appears next to a test statistic (r, t, F, X 2, Beta, etc.), the statistic is sufficiently large to overcome the null hypothesis. [test statistic]* =.05 (5 chances in 100 that the null hypo. is true) [test statistic]** =.01 (1 chance in 100 the null hypo. is true) [test statistic]*** =.001 (1 chance in 1,000 the null hypo is true) Null Hypothesis

8 Tests of Significance The specific test to be used depends on the level of measurement of the variables All variables categorical: Chi-Square (X 2 ) Independent variable categorical, dependent variable continuous: T-test and ANOVA All variables continuous: Correlation and regression Tests of significance yield a statistic, which is the ratio of: Systematic effect T he “real” influence of the independent variables STATISTIC = (e.g., r, Chi-Square, Chance/random effect “t”, z, “F”, Beta) The apparent influence, actually produced by chance The LARGER the numerator, the smaller the denominator, the larger the statistic As statistics get bigger, the probability that the null hypothesis is true falls When the probability that the null hypothesis is true falls to five chances in hundred (.05) or less, the relationship between the variables is deemed “statistically significant”

9 Estimating the error between sample and population statistics Standard error of the mean –An estimate of the amount of the (chance) error we can expect between the means of repeated samples and the population “parameter” s (standard deviation of our sample) S x = _____ ___  n-1 (sq. root of sample size minus 1) –We can’t use s without adjustment because the standard deviation of a sample contains real (systematic) differences in addition to differences that are due to error

10 Confidence limit: The range of values within which we are confident the population mean will fall Center point for estimating this range is the mean of a single, randomly drawn sample cl =  x  z (S x ) How “confident” must we be? In social science research, we do not want to exceed a chance of more than 5 in 100 of being wrong –z-scores represent an actual value on the variable being measured – the more extreme the value, the more extreme the z-score –Cases with extreme values are so rare that we could say they do not “belong” to the population from which the sample was drawn Confidence Limit and z scores

11 95% of the area under a standard normal curve is in the range 1.96z to +1.96z. So, for z, we usually use 1.96 Since the means of repeated samples are normally distributed around the population mean, we can use the normal curve and z scores to find the confidence limits

12 Class exercise – confidence limit Using data from the offender sample drawn earlier, calculate a confidence limit into which the population parameter will fall 95 out of 100 times s Sx = _____ ___  n-1 cl =  x  z (Sx)

13 Police recruit IQ test Sample of 100 recruits tested for IQ ( it is well established that IQ scores are normally distributed in the general population ).  x = 120 n = 100 s = 16 Standard error of the mean: s S x = --------- ___  n-1 Confidence limit: cl =  x  z (S x ) There is a 95% probability that the mean IQ of the population from which this sample was drawn falls between these scores: _______ and ________ lower limit upper limit

14  x = 120 n = 100 s = 16 s 16 S x = _____ = ________ = 1.61 ___ ____  n-1  99 cl = 120  1.96 (1.61) 95% probability that, based on the statistics of this sample, the mean IQ of the population from which the sample was drawn falls between 116.84 and 123.16


Download ppt "Inferential Statistics Introduction. If both variables are categorical, build tables... Convention: Each value of the independent (causal) variable has."

Similar presentations


Ads by Google