James R. Stacks, Ph.D. Introduction to Multiple Regression The best way to have a good idea is to have lots of ideas Linus.

Slides:



Advertisements
Similar presentations
Chapter 4 Systems of Linear Equations; Matrices
Advertisements

Chapter 4 Systems of Linear Equations; Matrices
TH EDITION LIAL HORNSBY SCHNEIDER COLLEGE ALGEBRA.
Matrices: Inverse Matrix
Chapter 2 Matrices Finite Mathematics & Its Applications, 11/e by Goldstein/Schneider/Siegel Copyright © 2014 Pearson Education, Inc.
Solving systems using matrices
Matrices and Systems of Equations
Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues
Review of Matrix Algebra
Chapter 1 Systems of Linear Equations
Matrix Algebra THE INVERSE OF A MATRIX © 2012 Pearson Education, Inc.
10.1 Gaussian Elimination Method
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 86 Chapter 2 Matrices.
LINEAR EQUATION IN TWO VARIABLES. System of equations or simultaneous equations – System of equations or simultaneous equations – A pair of linear equations.
Linear regression models in matrix terms. The regression function in matrix terms.
LIAL HORNSBY SCHNEIDER
Relationships Among Variables
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2011 Hawkes Learning Systems. All rights reserved. Hawkes Learning Systems College Algebra.
5  Systems of Linear Equations: ✦ An Introduction ✦ Unique Solutions ✦ Underdetermined and Overdetermined Systems  Matrices  Multiplication of Matrices.
Multivariate Linear Systems and Row Operations.
259 Lecture 14 Elementary Matrix Theory. 2 Matrix Definition  A matrix is a rectangular array of elements (usually numbers) written in rows and columns.
Chapter one Linear Equations
1 1.1 © 2012 Pearson Education, Inc. Linear Equations in Linear Algebra SYSTEMS OF LINEAR EQUATIONS.
Boyce/DiPrima 9th ed, Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues Elementary Differential Equations and Boundary Value Problems,
Systems and Matrices (Chapter5)
 Row and Reduced Row Echelon  Elementary Matrices.
Rev.S08 MAC 1140 Module 10 System of Equations and Inequalities II.
Copyright © 2011 Pearson, Inc. 7.3 Multivariate Linear Systems and Row Operations.
Slide Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley.
Solving Systems of Equations by matrices
Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall.
MATH 250 Linear Equations and Matrices
Some matrix stuff.
Copyright © 2013, 2009, 2005 Pearson Education, Inc. 1 5 Systems and Matrices Copyright © 2013, 2009, 2005 Pearson Education, Inc.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
Copyright © 2013, 2009, 2005 Pearson Education, Inc. 1 5 Systems and Matrices Copyright © 2013, 2009, 2005 Pearson Education, Inc.
Unit 3: Matrices.
Multivariate Statistics Matrix Algebra I W. M. van der Veld University of Amsterdam.
GG 313 Geological Data Analysis Lecture 13 Solution of Simultaneous Equations October 4, 2005.
Linear algebra: matrix Eigen-value Problems Eng. Hassan S. Migdadi Part 1.
Copyright © 2013, 2009, 2005 Pearson Education, Inc. 1 4 Inverse, Exponential, and Logarithmic Functions Copyright © 2013, 2009, 2005 Pearson Education,
4.5 Inverse of a Square Matrix
Copyright © 2011 Pearson Education, Inc. Solving Linear Systems Using Matrices Section 6.1 Matrices and Determinants.
Chapter 8 Matrices and Determinants Matrix Solutions to Linear Systems.
College Algebra Sixth Edition James Stewart Lothar Redlin Saleem Watson.
Slide Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley A set of equations is called a system of equations. The solution.
Learning Objectives for Section 4.5 Inverse of a Square Matrix
Chapter 1 Systems of Linear Equations Linear Algebra.
Matrices Digital Lesson. Copyright © by Houghton Mifflin Company, Inc. All rights reserved. 2 A matrix is a rectangular array of real numbers. Each entry.
TH EDITION LIAL HORNSBY SCHNEIDER COLLEGE ALGEBRA.
Copyright ©2015 Pearson Education, Inc. All rights reserved.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Slide Copyright © 2009 Pearson Education, Inc. 7.4 Solving Systems of Equations by Using Matrices.
Unit 3: Matrices. Matrix: A rectangular arrangement of data into rows and columns, identified by capital letters. Matrix Dimensions: Number of rows, m,
1 SYSTEM OF LINEAR EQUATIONS BASE OF VECTOR SPACE.
Section 6.2 Solving Linear Equations Math in Our World.
10.4 Matrix Algebra. 1. Matrix Notation A matrix is an array of numbers. Definition Definition: The Dimension of a matrix is m x n “m by n” where m =
Chapter 4 Systems of Linear Equations; Matrices
College Algebra Chapter 6 Matrices and Determinants and Applications
Linear Equations in Linear Algebra
Solving Systems of Equations Using Matrices
Copyright 2013, 2009, 2005, 2002 Pearson, Education, Inc.
5 Systems of Linear Equations and Matrices
Solving One-Step Equations
Unit 3: Matrices
Linear Equations in Linear Algebra
Solving Equations Using Multiplication and Division
Matrix Algebra THE INVERSE OF A MATRIX © 2012 Pearson Education, Inc.
Chapter 4 Systems of Linear Equations; Matrices
Presentation transcript:

James R. Stacks, Ph.D. Introduction to Multiple Regression The best way to have a good idea is to have lots of ideas Linus Pauling

Z ’ c =   Z p1 +   Z p2 +   Z p3 Standardized form of a regression equation with three predictor variables

Z ’ c =   Z p1 +   Z p2 +   Z p3 Predictor variables (standardized z scores)

Z ’ c =   Z p1 +   Z p2 +   Z p3 Standardized regression coefficients

Z ’ c =   Z p1 +   Z p2 +   Z p3 Predicted criterion score (z c – z e )

Z ’ c =   Z p1 +   Z p2 +   Z p3 Standardized regression coefficients Predictor variables (standardized z scores) Predicted criterion score (z c – z e )

Z ’ c =   Z p1 +   Z p2 +   Z p3 Predicted criterion score (z c – z e )

Z ’ c =   Z p1 +   Z p2 +   Z p3 Predicted criterion score (z c – z e ) Recall that the predicted criterion score is the is the actual criterion score minus the error Z c =   Z p1 +   Z p2 +   Z p3 + Z e

Recall that multiplication of an entire equation by any value results in an equivalent equation: y=bx is the same as yx = bxx or as yx = bx 2

The following demonstration of solving for standardized regression coefficients is taken largely from: Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.

Z c Z p3 =   Z p1 Z p3 +   Z p2 Z p3 +   Z p3 Z p3 + Z e Z p3 Z c Z p2 =   Z p1 Z p2 +   Z p2 Z p2 +   Z p3 Z p2 + Z e Z p2 Z c Z p1 =   Z p1 Z p1 +   Z p2 Z p1 +   Z p3 Z p1 + Z e Z p1 Let’s write three equivalent forms of the previous multiple regression equation by multiplying the original equation by each of the three predictor variables: (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

Now notice all the zz cross products in the equations. Recall that the expected (mean) cross product is something we are familiar with. The unbiased estimate of the cross product for paired z values is: E(cross product) =  zz)/(n-1), or, Pearson r ! Z c Z p1 =   Z p1 Z p1 +   Z p2 Z p1 +   Z p3 Z p1 + Z e Z p1 Z c Z p2 =   Z p1 Z p2 +   Z p2 Z p2 +   Z p3 Z p2 + Z e Z p2 Z c Z p3 =   Z p1 Z p3 +   Z p2 Z p3 +   Z p3 Z p3 + Z e Z p3 (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

The Pearson product-moment correlation coefficient ( written as r for sample estimate,  for parameter )  Z a Z b n-1 i = 1 n r = Where z a and z b are z scores for each person on some measure a and some measure b, and n is the number of people

r c p1 =   r  p1 p1 +   r p2 p1 +   r p3 p1 + r e p1 r c p2 =   r  p1 p2 +   r p2 p2 +   r p3 p2 + r e p2 r c p3 =   r  p1 p3 +   r p2 p3 +   r p3 p3 + r e p3 So, I could just as easily write: (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

r c p1 =   r  p1 p1 +   r p2 p1 +   r p3 p1 + r e p1 r c p2 =   r  p1 p2 +   r p2 p2 +   r p3 p2 + r e p2 r c p3 =   r  p1 p3 +   r p2 p3 +   r p3 p3 + r e p3 Now, let’s look at some interesting things about the correlation coefficients we have substituted Correlations of variables with themselves are necessarily unity, So the red values are 1 In regression, error by definition is the variance which does not correlate with any variable, thus the blue values are necessarily 0 (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

r c p1 =   (1) +   r p2 p1 +   r p3 p1 r c p2 =   r  p1 p2 +   (1) +   r p3 p2 r c p3 =   r  p1 p3 +   r p2 p3 +   (1) The above system can be written in matrix form: r c p1 r c p2 r c p3 =   (1) +   r p2 p1 +   r p3 p1   r  p1 p2 +   (1) +   r p3 p2   r  p1 p3 +   r p2 p3 +   (1) (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

r c p1 r c p2 r c p3 =   (1) +   r p2 p1 +   r p3 p1   r  p1 p2 +   (1) +   r p3 p2   r  p1 p3 +   r p2 p3 +   (1) Note that the matrix on the right side above is a vector, and it is a product of a correlation matrix of the predictor variables and a  vector. r c p1 r c p2 r c p3 =    (1)  r p2 p1  r p3 p1 r  p1 p2  (1)  r p3 p2 r  p1 p3  r p2 p3  (1) (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

r c p1 r c p2 r c p3 =    1  r p2 p1  r p3 p1 r  p1 p2  1  r p3 p2 r  p1 p3  r p2 p3  1 The moral of this story is: assuming all the Pearson correlations among variables are known (they are easily calculated), we can use the equation above to solve for the  vector, which is the standardized regression coefficients. Z ’ c =   Z p1 +   Z p2 +   Z p3 (Maruyama, Geoffrey M. (1998). Basics of structural equation modeling. Thousand Oaks, CA: Sage Publications, Inc.)

r c p1 r c p2 r c p3 =    1  r p2 p1  r p3 p1 r  p1 p2  1  r p3 p2 r  p1 p3  r p2 p3  1 This is a matrix equation which can be symbolized as: R iy = R ii B i From algebra, such an equation can obviously be solved for B i by dividing both sides by R ii, but there is no such thing as division in matrix math The matrix notation used here corresponds to your text: Tabachnik, Barbara G. & Fidell, Linda S. (2001). Using multivariate statistics., 4 th Edition. Needham Heights, MA: Allyn & Bacon

What is necessary to accomplish the same goal is to multiply both sides of the equation by the inverse of R ii, written as R ii -1. R ii -1 R iy = R ii -1 R ii B i therefore R ii -1 R iy = B i If you have studied the appendix assigned on matrix algebra,you know that, while matrix multiplication is quite simple, matrix inversion is a real chore! The matrix notation used here corresponds to your text: Tabachnik, Barbara G. & Fidell, Linda S. (2001). Using multivariate statistics., 4 th Edition. Needham Heights, MA: Allyn & Bacon

r c p1 r c p2 r c p3 =    1  r p2 p1  r p3 p1 r  p1 p2  1  r p3 p2 r  p1 p3  r p2 p3  1 To get the solution we must find the inverse of the green shaded matrix R ii in order to get R ii -1 for the equation : R ii -1 R iy = B i R iy = R ii B i The matrix notation used here corresponds to your text: Tabachnik, Barbara G. & Fidell, Linda S. (2001). Using multivariate statistics., 4 th Edition. Needham Heights, MA: Allyn & Bacon

The following method of inverting a matrix is taken largely from: Swokowski, Earl W. (1979) Fundamentals of College Algebra. Boston, MA: Prindle, Weber & Scmidt

The first step is to form a matrix which has the same number of rows as the original correlation matrix of predictors, but has twice as many columns. The original predictor correlations are placed in the left half, and an equal order identity matrix is place in the right half: (Predictor correlations) (Identity matrix) (Swokowski, Earl W. (1979) Fundamentals of College Algebra. Boston, MA: Prindle, Weber & Scmidt)

Though a series of calculations called elementary row transformations, the goal is to change all the numbers in the matrix so that the identity matrix is on the left, and a new matrix is on the right: Identity Matrix Inverse Matrix (Swokowski, Earl W. (1979) Fundamentals of College Algebra. Boston, MA: Prindle, Weber & Scmidt)

Swokowski, Earl W. (1979) Fundamentals of College Algebra. Boston, MA: Prindle, Weber & Scmidt “ MATRIX ROW TRANSFORMATION THEOREM Given a matrix of a system of linear equations, each of the following transformations results in a matrix of an equivalent system of linear equations: (i) (i)Interchanging any two rows (ii) (ii)Multiplying all of the elements in a row by the same nonzero real number k. (iii) (iii)Adding to the elements in a row k times the corresponding elements of any other row, where k is any real number. “

1 st transformation: a 2j = a 2j + (-.488) a 1j

2nd transformation: a 3j = a 3j + (-.354) a 1j

3rd transformation: a 2j = a 2j. 1 /

4th transformation: a 3j = a 3j + ( ) a 2j

5th transformation: a 3j = a 3j. 1 /

6th transformation: a 1j = a 1j + (-.488) a 2j

7th transformation: a 1j = a 1j + ( ) a 3j

8th transformation: a 2j = a 2j + ( ) a 3j

Original matrix on left Inverted matrix on right

=          beta vector inverse of predictor correlations predictor/criterion correlations R ii -1 R iy = B i

OUR CALCULATIONS VALUES FROM SPSS          The difference has to do with rounding error. There are so many transformations in matrix math that all computations must be carried out with many, many significant figures, because the errors accumulate. I only used what was visible in my calculator. Good matrix software should use much more precision. This is a relatively brief equation to solve. Imagine the error that can accumulate with hundreds of matrix transformations. This is a very important point, and one should always be certain the software is using the appropriate degree of precision.,

Z ’ c =  -  255  Z p1 +  872  Z p2 +  149  Z p3 The regression equation can then be written: