Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.

Slides:



Advertisements
Similar presentations
Chap 12-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 12 Simple Regression Statistics for Business and Economics 6.
Advertisements

1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Regression Greg C Elvers.
Regression, Correlation. Research Theoretical empirical Usually combination of the two.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
1 Review of Correlation A correlation coefficient measures the strength of a linear relation between two measurement variables. The measure is based on.
Chapter Topics Types of Regression Models
Statistics and Quantitative Analysis U4320
Simple Linear Regression Analysis
SIMPLE LINEAR REGRESSION
Ch. 14: The Multiple Regression Model building
Chapter 14 Introduction to Linear Regression and Correlation Analysis
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Correlation and Regression Analysis
Leon-Guerrero and Frankfort-Nachmias,
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Descriptive measures of the strength of a linear association r-squared and the (Pearson) correlation coefficient r.
Relationships Among Variables
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Correlation & Regression
Correlation and Regression
Leedy and Ormrod Ch. 11 Gray Ch. 14
Chapter 8: Bivariate Regression and Correlation
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Introduction to Linear Regression and Correlation Analysis
Chapter 11 Simple Regression
Correlation and Linear Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Examining Relationships in Quantitative Research
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
Intro to Regression POL 242. Summary Regression is the process by which we fit a line to depict the relationship between two variables (usually both interval.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 16 Data Analysis: Testing for Associations.
Regression Analysis Relationship with one independent variable.
Political Science 30: Political Inquiry. Linear Regression II: Making Sense of Regression Results Interpreting SPSS regression output Coefficients for.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Correlation & Regression Analysis
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
Lesson 14 - R Chapter 14 Review. Objectives Summarize the chapter Define the vocabulary used Complete all objectives Successfully answer any of the review.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
4 basic analytical tasks in statistics: 1)Comparing scores across groups  look for differences in means 2)Cross-tabulating categoric variables  look.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
The simple linear regression model and parameter estimation
Lecture 10 Regression Analysis
REGRESSION G&W p
Bivariate & Multivariate Regression Analysis
Introduction to Regression Analysis
Product moment correlation
Regression Part II.
Presentation transcript:

Multiple Independent Variables POLS 300 Butz

Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality Need for techniques that allow the research to control for other independent variables

Multivariate Analysis Employed to see how large sets of variables are interrelated. Idea is that if one can find a relationship between x and y after accounting for other variables (w and z) we may be able to make “causal inference”.

Multivariate Analysis We know that both X and Y both may be caused by Z, spurious relationship. Multivariate Analysis allows for the inclusions of other variables and to test if there is still a relationship between X and Y.

Multivariate Analysis Must ask if the possibility of a third variable (and maybe others) is the “true” cause of both the IV and DV Experimental analyses “prove” causation but only in Laboratory Setting…must use Multivariate Statistical Analyses in “real- world” Need to “Control” or “hold constant” other variables to isolate the effect of IV on DV!

Controlling for Other Independent Variables Multivariate Crosstabulation – evaluate bivariate relationship within subsets of sample defined by different categories of third variable (“control by grouping”) At what level(s) of measurement would we use Multivariate Crosstabulation??

Multivariate Crosstabulation Control by grouping: group the observations according to their values on the third variable and… then observe the original relationship within each of these groups. P. 407/506 – Spending Attitudes and Voting…controlling for Income! – spurious Occupational Status and Voter Turnout P. 411/510…control for “education”!

Quick Review: Regression In general, the goal of linear regression is to find the line that best predicts Y from X. Linear regression does this by estimating a line that minimizes the sum of the squared errors from the line Minimizing the vertical distances of the data points from the line.

Regression vs. Correlation The purpose of regression analysis is to determine exactly what the line is (i.e. to estimate the equation for the line) The regression line represents predicted values of Y based on the values of X

Equation for a Line (Perfect Linear Relationship) Y i = a + BX i a = Intercept, or Constant = The value of Y when X = 0 B = Slope coefficient = The change (+ or ‑ ) in Y given a one unit change in X

Slope Y i = a + BX i B = Slope coefficient If B is positive than you have a positive relationship. If it is negative you have a negative relationship. The larger the value of B the more steep the slope of the line…Greater (more dramatic) change in Y for a unit change in X General Interpretation: For one unit change in X, we expect a B change in Y on average.

Calculating the Regression Equation For “Threat Hypothesis” The estimated regression equation is: E(welfare benefit1995) = [(-6.292) * %black(1995)] Number of obs = 50 F( 1, 64) = Prob < = R-squared = welfare1995 | Coef. Std. Err. t P< [95% Conf. Interval] Black1995(b)| _cons(a)|

Regression Example: “Threat Hypothesis” To generate a predicted value for various % of AA in 1995, we could simply plug in the appropriate X values and solve for Y. 10% E(welfare benefit1995) = [(-6.292) * 10] = $ % E(welfare benefit1995) = [(-6.292) * 20] = $ % E(welfare benefit1995) = [(-6.292) * 30] = $234.09

Regression Analysis and Statistical Significance Testing for statistical significance for the slope –The p-value - probability of observing a sample slope value (Beta Coefficent) at least as large as the one we are observing in our sample IF THE NULL HYPOTHESIS IS TRUE –P-values closer to 0 suggest the null hypothesis is less likely to be true (P <.05 usually the threshold for statistical significance) –Based on t-value…(Beta/S.E.) = t

Multiple Regression At what level(s) of measurement would we employ multiple regression??? Interval and Ratio DVs Now working with a new model: Y i = a  b  X  i  b  X 2i  b k X ki  e i

Multiple Regression Y i = a  b  X  i  b  X 2i  b k X ki  e i b are “Partial” slope coefficients. a is the Y-Intercept. e is the Error Term.

Slope Coefficients Slope coefficients are now Partial Slope Coefficients, although we still refer to them generally as slope coefficients. They have a new interpretation: “The expected change in Y given a one ‑ unit change in X1, holding all other variables constant”

Multiple Regression By “holding constant” all other X’s, we are therefore “controlling for” all other X’s, and thus isolating the “independent effect” of the variable of interest. “Holding Constant” – group observations according to levels of X2, X3, ect…then look at impact of X1 on Y! This is what Multiple Regression is doing in practice!!! Make everyone “equal” in terms of “control” variable then examine the impact of X1 on Y!

“Holding Constant” other IVs Income (Y) = Education (X1); Seniority (X2) Look at relationship between Seniority and Income WITHIN different levels of education!!! “Holding Education Constant” Look at relationship between Education and Income WITHIN different levels of Senority!!! “Holding Seniority Constant”

The Intercept Y i = a  b  X  i  b  X 2i  b k X ki  e i Y-Intercept (Constant) value…(a)…is now the expected value of Y when ALL the Independent Variables are set to 0.

Testing for Statistical Significance Proceeds as before – a probability that the null hypothesis holds (p-value) is generated for each sample slope coefficient Based on “t-value” (Beta/ S.E.) And Degrees of Freedom!

Fit of the Regression R-squared value – the proportion of variation in the dependent variable explained by ALL of the independent variables combined TSS – ResSS/ TSS… “Explained Variation in DV divided by Total Variation in DV”

R-square R-square ranges from 0 to 1. 0 is no relationship. 1 is a prefect relationship…IVs explain 100% of the variance in the DV.

R-square Doesn’t tell us WHY the dependent variable varies or explains the results….This is why we need Theory!!! Simply a measure of how well your model fits the dependent variable. How well are the Xs predicting Y! How much variation in Y is explained by Xs!

Multiple Regression -Y= Income in dollars X1= Education in years X2= Seniority in years Y= a + b1(education) + b2(Seniority) + e

Example Y= X X 2 + e - Both Coefficients are statistically significant at the P <.05 Level… Because of the positive Beta…expected change in Income (Y) given a one ‑ unit increase in Education is +$432, holding seniority in years constant.

Predicted Values Lets predict someone with 10 years of education and 5 years of seniority. Y= X X 2 +e = (10)+281(5) = Predicted value of Y for this case is $11,391.

R-squared r-squared for this model is.56. Education and Seniority explain 56% of the variation in income.