Bivariate Analyses.

Slides:



Advertisements
Similar presentations
Simple Linear Regression and Correlation by Asst. Prof. Dr. Min Aung.
Advertisements

Review ? ? ? I am examining differences in the mean between groups
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Hypothesis Testing IV Chi Square.
Describing Relationships Using Correlation and Regression
t-Tests Overview of t-Tests How a t-Test Works How a t-Test Works Single-Sample t Single-Sample t Independent Samples t Independent Samples t Paired.
T-Tests.
PSY 307 – Statistics for the Behavioral Sciences
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Chapter18 Determining and Interpreting Associations Among Variables.
The Simple Regression Model
Session 7.1 Bivariate Data Analysis
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 6: Correlation.
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Social Research Methods
Chapter 9 - Lecture 2 Computing the analysis of variance for simple experiments (single factor, unrelated groups experiments).
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Linear Regression/Correlation
Relationships Among Variables
(and a bit on regression)
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Inferential Statistics
Correlation and Regression A BRIEF overview Correlation Coefficients l Continuous IV & DV l or dichotomous variables (code as 0-1) n mean interpreted.
Leedy and Ormrod Ch. 11 Gray Ch. 14
Lecture 16 Correlation and Coefficient of Correlation
Statistics for the Social Sciences Psychology 340 Fall 2013 Thursday, November 21 Review for Exam #4.
AM Recitation 2/10/11.
LIS 570 Summarising and presenting data - Univariate analysis continued Bivariate analysis.
Learning Objective Chapter 14 Correlation and Regression Analysis CHAPTER fourteen Correlation and Regression Analysis Copyright © 2000 by John Wiley &
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Data Analysis (continued). Analyzing the Results of Research Investigations Two basic ways of describing the results Two basic ways of describing the.
Hypothesis of Association: Correlation
Correlations. Outline What is a correlation? What is a correlation? What is a scatterplot? What is a scatterplot? What type of information is provided.
POSC 202A: Lecture 12/10 Announcements: “Lab” Tomorrow; Final ed out tomorrow or Friday. I will make it due Wed, 5pm. Aren’t I tender? Lecture: Substantive.
Examining Relationships in Quantitative Research
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
TYPES OF STATISTICAL METHODS USED IN PSYCHOLOGY Statistics.
URBP 204A QUANTITATIVE METHODS I Statistical Analysis Lecture IV Gregory Newmark San Jose State University (This lecture is based on Chapters 5,12,13,
Correlation & Regression Chapter 15. Correlation It is a statistical technique that is used to measure and describe a relationship between two variables.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
Chapter Thirteen Copyright © 2006 John Wiley & Sons, Inc. Bivariate Correlation and Regression.
Statistics in IB Biology Error bars, standard deviation, t-test and more.
Correlations. Distinguishing Characteristics of Correlation Correlational procedures involve one sample containing all pairs of X and Y scores Correlational.
Statistical Analysis. Z-scores A z-score = how many standard deviations a score is from the mean (-/+) Z-scores thus allow us to transform the mean to.
Chapter Eight: Using Statistics to Answer Questions.
Correlation & Regression Analysis
Statistics. Descriptive Statistics Organize & summarize data (ex: central tendency & variability.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Advanced Statistical Methods: Continuous Variables REVIEW Dr. Irina Tomescu-Dubrow.
T-tests Chi-square Seminar 7. The previous week… We examined the z-test and one-sample t-test. Psychologists seldom use them, but they are useful to understand.
Copyright © 2008 by Nelson, a division of Thomson Canada Limited Chapter 18 Part 5 Analysis and Interpretation of Data DIFFERENCES BETWEEN GROUPS AND RELATIONSHIPS.
Bivariate Association. Introduction This chapter is about measures of association This chapter is about measures of association These are designed to.
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Qualitative data – tests of association
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
Social Research Methods
Summarising and presenting data - Bivariate analysis
Correlation and Regression
BIVARIATE ANALYSIS: Measures of Association Between Two Variables
4/4/2019 Correlations.
15.1 The Role of Statistics in the Research Process
BIVARIATE ANALYSIS: Measures of Association Between Two Variables
UNIT V CHISQUARE DISTRIBUTION
Review I am examining differences in the mean between groups How many independent variables? OneMore than one How many groups? Two More than two ?? ?
Warsaw Summer School 2017, OSU Study Abroad Program
Presentation transcript:

Bivariate Analyses

Bivariate Procedures I Overview Chi-square test T-test Correlation

Chi-Square Test Relationships between nominal variables Types: 2x2 chi-square Gender by Political Party 2x3 chi-square Gender by Dosage (Hi vs. Med. Vs. Low)

Starting Point: The Crosstab Table Example: Gender (IV) Males Females Democrat 1 20 Party (DV) Republican 10 2 Total 11 22

Column Percentages Gender (IV) Males Females Democrat 9% 91% Party (DV) Republican 91% 9% Total 100% 100%

Row Percentages Gender (IV) Males Females Total Democrat 5% 95% 100% Party (DV) Republican 83% 17% 100%

Full Crosstab Table Males Females Total Democrat 1 20 21 5% 95% 5% 95% 9% 91% 64% Republican 10 2 12 83% 17% 91% 9% 36% Total 11 22 33 33% 67% 100%

Research Question and Hypothesis Is gender related to party affiliation? Hypothesis: Men are more likely than women to be Republicans Null hypothesis: There is no relation between gender and party

Testing the Hypothesis Eyeballing the table: Seems to be a relationship Is it significant? Or, could it be just a chance finding? Logic: Is the finding different enough from the null? Chi-square answers this question What factors would it take into account?

Factors Taken into Consideration 1. Magnitude of the difference 2. Sample size Biased coin example Magnitude of difference: 60% heads vs. 99% heads Sample size: 10 flips vs. 100 flips vs. 1 million flips

Chi-square Chi-Square starts with the frequencies: Compare observed frequencies with frequencies we expect under the null hypothesis

What would the Frequencies be if there was No Relationship? Males Females Total Democrat 21 Republican 12 Total 11 22 33

Expected Frequencies (Null) Males Females Total Democrat 7 14 21 Republican 4 8 12 Total 11 22 33

Comparing the Observed and Expected Cell Frequencies Formula:

Calculating the Expected Frequency Simple formula for expected cell frequencies Row total x column total / Total N 21 x 11 / 33 = 7 21 x 22 / 33 = 14 12 x 11 / 33 = 4 12 x 22 / 33 = 8

Observed and Expected Cell Frequencies Males Females Total Democrat 1 7 20 14 21 Republican 10 4 2 8 12 Total 11 22 33

Plugging into the Formula O - E Square Square/E Cell A = 1-7 = -6 36 36/7 = 5.1 Cell B = 20-14 = 6 36 36/14 = 2.6 Cell C = 10-4 = 6 36 36/4 = 9 Cell D = 2-8 = -6 36 36/8 = 4.5 Sum = 21.2 Chi-square = 21.2

Is the chi-square significant? Significance of the chi-square: Great differences between observed and expected lead to bigger chi-square How big does it have to be for significance? Depends on the “degrees of freedom” Formula for degrees of freedom: (Rows – 1) x (Columns – 1)

Chi-square Degrees of Freedom 2 x 2 chi-square = 1 3 x 3 = ? 4 x 3 = ?

Chi-square Critical Values df P = 0.05 P = 0.01 P = 0.001 1 3.84 6.64 10.83 2 5.99 9.21 13.82 3 7.82 11.35 16.27 4 9.49 13.28 18.47 5 11.07 15.09 20.52 6 12.59 16.81 22.46 7 14.07 18.48 24.32 8 15.51 20.09 26.13 9 16.92 21.67 27.88 10 18.31 23.21 29.59 * If chi-square is > than critical value, relationship is significant

Chi-Square Computer Printout

Chi-Square Computer Printout

Multiple Chi-square Exact same procedure as 2 variable X2 Used for more than 2 variables E.g., 2 x 2 x 2 X2 Gender x Hair color x eye color

Multiple chi-square example

Multiple chi-square example

The T-test Groups T-test Pairs T-test Comparing the means of two nominal groups E.g., Gender and IQ E.g., Experimental vs. Control group Pairs T-test Comparing the means of two variables Comparing the mean of a variable at two points in time

Logic of the T-test A T-test considers three things: 1. The group means 2. The dispersion of individual scores around the mean for each group (sd) 3. The size of the groups

Difference in the Means The farther apart the means are: The more confident we are that the two group means are different Distance between the means goes in the numerator of the t-test formula

Why Dispersion Matters Small variances Large variances

Size of the Groups Larger groups mean that we are more confident in the group means IQ example: Women: mean = 103 Men: mean = 97 If our sample was 5 men and 5 women, we are not that confident If our sample was 5 million men and 5 million women, we are much more confident

The four t-test formulae 1. Matched samples with unequal variances 2. Matched samples with equal variances 3. Independent samples with unequal variances 4. Independent samples with equal variances

All four formulae have the same Numerator X1 - X2 (group one mean - group two mean) What differentiates the four formulae is their denominator denominator is “standard error of the difference of the means” each formula has a different standard error

Independent sample with unequal variances formula Standard error formula (denominator):

T-test Value Look up the T-value in a T-table (use absolute value ) First determine the degrees of freedom ex. df = (N1 - 1) + (N2 - 1) 40 + 30 = 70 For 70 df at the .05 level =1.67 ex. 5.91 > 1.67: Reject the null (means are different)

Groups t-test printout example

Pairs t-test example

Pearson Correlation Coefficient (r ) Characteristics of correlational relationships: 1. Strength 2. Significance 3. Directionality 4. Curvilinearity

Strength of Correlation: Strong, weak and non-relationships Nature of such relations can be observed in scatter diagrams Scatter diagram One variable on x axis and the other on the y-axis of a graph Plot each case according to its x and y values

Scatterplot: Strong relationship B O K R E A D I N G Years of Education

Scatterplot: Weak relationship M E Years of Education

Scatterplot: No relationship Years of Education

Strength increases… As the points more closely conform to a straight line Drawing the best fitting line between the points: “the regression line” Minimizes the distance of the points from the line: “least squares” Minimizing the deviations from the line

Significance of the relationship Whether we are confident that an observed relationship is “real” or due to chance What is the likelihood of getting results like this if the null hypothesis were true? Compare observed results to expected under the null If less than 5% chance, reject the null hypothesis

Directionality of the relationship Correlational relationship can be positive or negative Positive relationship High scores on variable X are associated with high scores on variable Y Negative relationship High scores on variable X are associated with low scores on variable Y

Positive relationship example B O K R E A D I N G Years of Education

Negative relationship example C I L P E J U D Years of Education

Curvilinear relationships Positive and negative relationships are “straight-line” or “linear” relationships Relationships can also be strong and curvilinear too Points conform to a curved line

Curvilinear relationship example F A M I L Y S Z E SES

Curvilinear relationships Linear statistics (e.g. correlation coefficient, regression) can mask a significant curvilinear relationship Correlation coefficient would indicate no relationship

Pearson Correlation Coefficient Numerical expression of: Strength and Direction of straight-line relationship Varies between –1 and 1

Correlation coefficient outcomes -1 is a perfect negative relationship -.7 is a strong negative relationship -.4 is a moderate negative relationship -.1 is a weak negative relationship 0 is no relationship .1 is a weak positive relationship .4 is a moderate positive relationship .7 is a strong positive relationship 1 is a perfect positive relationship

Pearson’s r (correlation coefficient) Used for interval or ratio variables Reflects the extent to which cases have similar z-scores on variables X and Y Positive relationship—z-scores have the same sign Negative relationship—z-scores have the opposite sign

Positive relationship z-scores Person Xz Yz A 1.06 1.11 B .56 .65 C .03 -.01 D -.42 -.55 E -1.23 -1.09

Negative relationship z-scores Person Xz Yz A 1.06 -1.22 B .56 -.51 C .03 -.06 D -.42 .66 E -1.23 1.33

Conceptual formula for Pearson’s r Multiply each cases z-score Sum the products Divide by N

Significance of Pearson’s r Pearson’s r tells us the strength and direction Significance is determined by converting the r to a t ratio and looking it up in a t table Null: r = .00 How different is what we observe from null? Less than .05?

Computer Printout