Statistics for Linguistics Students Michaelmas 2004 Week 7 Bettina Braun www.phon.ox.ac.uk/~bettina/teaching.html.

Slides:



Advertisements
Similar presentations
Statistics for Linguistics Students Michaelmas 2004 Week 5 Bettina Braun
Advertisements

CORRELATION. Overview of Correlation u What is a Correlation? u Correlation Coefficients u Coefficient of Determination u Test for Significance u Correlation.
Correlation and regression
Inference for Regression
Statistics for Linguistics Students Michaelmas 2004 Week 6 Bettina Braun
Correlation Mechanics. Covariance The variance shared by two variables When X and Y move in the same direction (i.e. their deviations from the mean are.
PSY 307 – Statistics for the Behavioral Sciences
Correlation. Introduction Two meanings of correlation –Research design –Statistical Relationship –Scatterplots.
CORRELATION. Overview of Correlation u What is a Correlation? u Correlation Coefficients u Coefficient of Determination u Test for Significance u Correlation.
Correlation and Regression. Spearman's rank correlation An alternative to correlation that does not make so many assumptions Still measures the strength.
CJ 526 Statistical Analysis in Criminal Justice
Chi Square Test Dealing with categorical dependant variable.
Lecture 9: One Way ANOVA Between Subjects
ANOVA  Used to test difference of means between 3 or more groups. Assumptions: Independent samples Normal distribution Equal Variance.
Dr. Mario MazzocchiResearch Methods & Data Analysis1 Correlation and regression analysis Week 8 Research Methods & Data Analysis.
Inferential Statistics
Inferential Statistics: SPSS
Simple Covariation Focus is still on ‘Understanding the Variability” With Group Difference approaches, issue has been: Can group membership (based on ‘levels.
Correlation.
CJ 526 Statistical Analysis in Criminal Justice
Week 10 Chapter 10 - Hypothesis Testing III : The Analysis of Variance
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Hypothesis of Association: Correlation
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
Chapter 16 The Chi-Square Statistic
1 Review of ANOVA & Inferences About The Pearson Correlation Coefficient Heibatollah Baghi, and Mastee Badii.
By: Amani Albraikan.  Pearson r  Spearman rho  Linearity  Range restrictions  Outliers  Beware of spurious correlations….take care in interpretation.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
1 Inferences About The Pearson Correlation Coefficient.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
CHI SQUARE TESTS.
Political Science 30: Political Inquiry. Linear Regression II: Making Sense of Regression Results Interpreting SPSS regression output Coefficients for.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
The basic task of most research = Bivariate Analysis
The basic task of most research = Bivariate Analysis A.What does that involve?  Analyzing the interrelationship of 2 variables  Null hypothesis = independence.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Analysis of Variance ANOVA. Example Suppose that we have five educational levels in our population denoted by 1, 2, 3, 4, 5 We measure the hours per week.
Chi Square & Correlation
Copyright c 2001 The McGraw-Hill Companies, Inc.1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent variable.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent.
Chapter 13 Understanding research results: statistical inference.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Methods of Presenting and Interpreting Information Class 9.
Inference about the slope parameter and correlation
Inferential Statistics
Correlation analysis is undertaken to define the strength an direction of a linear relationship between two variables Two measurements are use to assess.
REGRESSION G&W p
Psych 706: stats II Class #4.
Dr. Siti Nor Binti Yaacob
Non-Parametric Tests 12/1.
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
B&A ; and REGRESSION - ANCOVA B&A ; and
Non-Parametric Tests 12/1.
Political Science 30: Political Inquiry
Making Use of Associations Tests
Non-Parametric Tests 12/6.
Non-Parametric Tests.
Data Analysis and Interpretation
Theme 7 Correlation.
Multiple logistic regression
Inferential Statistics
Some statistics questions answered:
Parametric versus Nonparametric (Chi-square)
Chapter 18: The Chi-Square Statistic
Exercise 1 Use Transform  Compute variable to calculate weight lost by each person Calculate the overall mean weight lost Calculate the means and standard.
InferentIal StatIstIcs
Inferential testing.
Presentation transcript:

Statistics for Linguistics Students Michaelmas 2004 Week 7 Bettina Braun

Overview Problems from last assignment Correlation analyses Repeated measures ANOVA –One-way (one IV) –Two-way (two IVs) Transformations

Chi-square using SPSS Organisation of data:

Chi-square using SPSS Where to find it…

Chi-square using SPSS How to interpret the output Table similar to ours Result: sign. interaction (x 2 =5.7, df=1, p=0.017

More on interactions NorthSouth Male Female NorthSouth No effect of region, nor gender, no interaction Effect of region and gender no interaction NorthSouth No effect of gender, effect of region, no interaction NorthSouth Effect of region and gender and interaction NorthSouth Effect of region and gender and interaction

Correlation analyses Often found in exploratory research –You do not test the effect of an independent variable on the dependent one –But see what relationships hold between two or more variables

Correlation coefficients Scatterplots helpful to see whether it is a linear relationship… r = -1 Neg. corr. r = 0 no corr. r = 1 pos. corr.

Bivariate correlation Do you expect a correlation between the two variables? Try “line-fitting” by eye ?

Pearson correlation T-test is used to test if corr. coefficient is different from 0 ( => data must be interval!) If not, use Spearmans correlation (non- parametric)

Pearson correlation Correlation coefficient –For interval data –For linear relationships r 2 is the proportion of variation of one variable that is “explained” by the other Note: even a highly significant correlation does not imply a causal relationship (e.g. There might be another variable influencing both!)

Repeated measures ANOVA Recall: –In between-subjects designs large individual differences –repeated measures (aka within-subjects) has all participants in all levels of all conditions Problems: –Practice effect (carry-over) effect

Missing data You need to have data for every subject in every condition If this is not the case, you cannot include this subject If your design becomes inbalanced by the exclusion of a subject, you should randomly exclude a subject from the other group as well (or run another subject for the group with the exclusion)

Requirements for repeated measures ANOVA Same as for between-subjects ANOVA You can have within- and between-subject factors (e.g. boys vs. girls, producing /a/ and /i/ and /u/) Covariates –factors that might have an effect on the within- subjects factor –Note: covariates can also be specified for between-subjects designs!

Covariates: example You want to study French skills when using 2 different text-books. Students are randomly assigned to 2 groups. If you have the IQ of these students, you can decrease the variability within the groups by using IQ as covariate Problem: if the covariate is correlated with between-groups factor as well, F-value might get smaller (less significant)! You can also assess interaction between covariates and between-groups factors (e.g. one textbook might be better suited for smart students)

One-way repeated measures ANOVA in SPSS Define new name and levels for within-subject factor

One-way repeated measures ANOVA in SPSS Factor-name Four levels of the within-subjects variable Enter between- subjects and covariates (if applicable)

Post-hoc tests for within-subjects variables SPSS does not allow you to do post-hoc tests for within-subjects variables Instead do “Contrasts” and define them as “Repeated” 2 1

Post-hoc tests for within-subjects variables You can also ask for a comparson of means

SPSS output: test of Sphericity Test for homgeneity of covariances among scores of within-subjecs factors Only calculated if variable has more than 2 levels If test is significant, you have to reject the null-hypothesis that the variances are homogenious

SPSS output: within-subjects contrasts Post-hoc test for within-subjects variables

3 x 3 designs 3 x 3 between subjects Factor B (between) B1B2B3 A1Group1Group2Group3 A2Group4Group5Group6 A3Group7Group8Group9

3 x 3 designs 3 x 3 within subjects Factor B (witin) B1B2B3 A1Group1 A2 A3 Group1

3 x 3 designs 3 x 3 mixed design Factor B (witin) B1B2B3 Factor A (between) A1Group1 A2 A3 Group1 Group2 Group3 Group2 Group3

Data transformation If you want to caculate an ANOVA but your interval data is not normally distributed (i.e. skewed) you can use mathematical transformations The type of transformation depends on the shape of the sample distribution NOTE: –After transforming data, check the resulting distribution again for normality! –Note that your data becomes ordinal by transforming it!! (but you can do an ANOVA with it)

What kind of tranformation? e.g. f(x) = x 1.5 e.g. f(x) = log(x) f(x) = atan(x) Transformation