Fundamental Statistics in Applied Linguistics Research Spring 2010 Weekend MA Program on Applied English Dr. Da-Fu Huang.

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Chapter 12 Simple Linear Regression
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Prediction with multiple variables Statistics for the Social Sciences Psychology 340 Spring 2010.
Statistics for the Social Sciences Psychology 340 Spring 2005 Prediction cont.
Statistics for the Social Sciences
Lecture 6: Multiple Regression
Examining Relationship of Variables  Response (dependent) variable - measures the outcome of a study.  Explanatory (Independent) variable - explains.
Chapter Topics Types of Regression Models
Linear Regression and Correlation Analysis
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
RESEARCH STATISTICS Jobayer Hossain Larry Holmes, Jr November 6, 2008 Examining Relationship of Variables.
Regression The basic problem Regression and Correlation Accuracy of prediction in regression Hypothesis testing Regression with multiple predictors.
Ch. 14: The Multiple Regression Model building
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Correlation and Regression Analysis
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Multiple Regression Dr. Andy Field.
Simple Linear Regression Analysis
Multiple Linear Regression Response Variable: Y Explanatory Variables: X 1,...,X k Model (Extension of Simple Regression): E(Y) =  +  1 X 1 +  +  k.
Advantages of Multivariate Analysis Close resemblance to how the researcher thinks. Close resemblance to how the researcher thinks. Easy visualisation.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Fundamental Statistics in Applied Linguistics Research Spring 2010 Weekend MA Program on Applied English Dr. Da-Fu Huang.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Regression Analysis Week 8 DIAGNOSTIC AND REMEDIAL MEASURES Residuals The main purpose examining residuals Diagnostic for Residuals Test involving residuals.
Lesson Multiple Regression Models. Objectives Obtain the correlation matrix Use technology to find a multiple regression equation Interpret the.
Regression Lesson 11. The General Linear Model n Relationship b/n predictor & outcome variables form straight line l Correlation, regression, t-tests,
Multiple Regression BPS chapter 28 © 2006 W.H. Freeman and Company.
Chapter 13 Multiple Regression
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
INDE 6335 ENGINEERING ADMINISTRATION SURVEY DESIGN Dr. Christopher A. Chung Dept. of Industrial Engineering.
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Examining Relationships in Quantitative Research
Ch14: Linear Least Squares 14.1: INTRO: Fitting a pth-order polynomial will require finding (p+1) coefficients from the data. Thus, a straight line (p=1)
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 10 th Edition.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Venn diagram shows (R 2 ) the amount of variance in Y that is explained by X. Unexplained Variance in Y. (1-R 2 ) =.36, 36% R 2 =.64 (64%)
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 9 l Simple Linear Regression 9.1 Simple Linear Regression 9.2 Scatter Diagram 9.3 Graphical.
Chapter 11 REGRESSION Multiple Regression  Uses  Explanation  Prediction.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Linear Regression.
Correlation, Bivariate Regression, and Multiple Regression
CHAPTER 3 Describing Relationships
Statistics for the Social Sciences
Kin 304 Regression Linear Regression Least Sum of Squares
Multiple Regression.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Chapter 12: Regression Diagnostics
BPK 304W Regression Linear Regression Least Sum of Squares
BPK 304W Correlation.
CHAPTER 29: Multiple Regression*
Multiple Regression A curvilinear relationship between one variable and the values of two or more other independent variables. Y = intercept + (slope1.
Correlation and Regression
Multiple Linear Regression
Presentation transcript:

Fundamental Statistics in Applied Linguistics Research Spring 2010 Weekend MA Program on Applied English Dr. Da-Fu Huang

6. Looking for groups of explanatory variables through multiple regression Explanatory variables vs. response variables MR examines whether the explanatory variables (EV) we’ve posited explain very much of what is going on in response variables (RV) Y = α + β 1 x i1 + … + β k x ik + error i TOEFL score = some constant number (the intercept; α ) + time spent on English per week ( β 1 ) + aptitude score ( β 2 ) + a number which fluctuates for each individual (the error) MR can also predicts how people in the future will score on the response variable

Venn diagram of regression variables Hours of study per week MLAT score Personality TOEFL score

The mathematical formula of a line Line equation: y = x slope = 0.5 intercept = 2 ‧ ‧ ‧ ‧ ‧ actual value Y 1 predicted value Y’ 1 Error or residual ‧ ‧ ‧ ‧ ‧ Regression line

The regression line ‧ ‧ ‧ ‧ ‧ Error or residual The best fitting line (closest to the data points) The least squares regression line (the line that minimizes the sum of the squared errors about the line; Σ(Y-Y’) 2 is minimized

6. Looking for groups of explanatory variables through multiple regression 6.1 Standard multiple regression (SMR) In SMR, the importance of the EV variable depends on how much it uniquely overlaps with the RV. SMR answers the two questions: What are the nature and size of the relationship between the RV and the set of EV? How much of the relationship is contributed uniquely by each EV?

Venn diagram of standard regression design Hours of study per week MLAT score Personality TOEFL score a b c d e

6. Looking for groups of explanatory variables through multiple regression 6.2 Sequential (Hierarchical) multiple regression (HMR) In HMR, all of the areas of the EV’s that overlap with the RV will be counted, but the way that they will be included depends on the order in which the researcher enters the variables into the equation The importance of any variable can be emphasized in HMR, depending on the order in which it is entered. If two variables overlap to a large degree, then entering one of them first will leave little room for explanation for the second variable HMR answers the question: Do the subsequent variables entered in each step add to the prediction of the RV after differences in the variables from the previous step have been eliminated?

Venn diagram of sequential regression design Hours of study per week MLAT score Personality TOEFL score a b c d e H  M  P

Assumptions for MR Table 7.1 (P184) Normal distribution Homogeneity of variances Linearity Multicollinearity (EV’s involved in the regression should not be highly intercorrelated)

6. Looking for groups of explanatory variables through multiple regression 6.4 Starting the MR (PP ) Analyze > Regression > Linear Put the RV in the box “Dependent” For Standard regression: put all EV into the “Independent” box with the Method set at “Enter” For sequential regression: put all EV’s into the “Independent” box with the Method set at “Enter”. Push the Next button after entering each one. Enter the EV in the order you want them into the regression equation. Open the buttons: Statistics, Plots, and Options

6. Looking for groups of explanatory variables through multiple regression 6.5 Regression output in SPSS Analyze > Regression > Linear

Regression Output

Regression Output (Standard)

Regression Output (Sequential)

Regression output (Standard)

Regression Output (Sequential)

Regression output (Standard) Under 5 Y= *EngProf *Anx +.46*Mid *EvaTch *Motiv T-test

Regression Output (Sequential) Y= (6.73)*EngProf + (5.56)*Motiv

Regression Output Check outliers

Regression Output: P-P plot for diagnosing normal distribution of data Check normality assumption Look at distribution of residuals, not individual variables

Regression Output: Plot of studentized residuals crossed with fitted values The shape should show a cloud of data scattered randomly Check homogeneity of variances

6. Looking for groups of explanatory variables through multiple regression 6.6 Reporting the results of regression analysis Correlations between the explanatory variables and the response variable Correlations among the explanatory variables Correlation matrix with r-value, p-value, and N Standard or sequential regression? R square or R square change for each step of the model Regression coefficients for all regression models (esp. unstandarized coefficients, labeled B, and the coefficient for the intercept, labeled “constant” in SPSS output) For standard regression, report the t-tests for the contribution of each variable to the model

6. Looking for groups of explanatory variables through multiple regression 6.6 Reporting the results of regression analysis The multiple correlation coefficient, R 2, expresses how much of the variable in scores of the response variable can be explained by the variance in the statistical explanatory variables The squared semipartial correlations ( sr 2 ) provides a way of assessing the unique contribution of each variable to the overall R. These numbers are already a percentage variance effect size (of the r family) Example reporting on Lafrance & Gottardo (2005): P198

Application activities (Q1-Q6): PP