MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.

Slides:



Advertisements
Similar presentations
All Possible Regressions and Statistics for Comparing Models
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Best subsets regression
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Statistics for Managers Using Microsoft® Excel 5th Edition
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Statistics for Managers Using Microsoft® Excel 5th Edition
Part I – MULTIVARIATE ANALYSIS C3 Multiple Linear Regression II © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
Correlation MARE 250 Dr. Jason Turner.
Lecture 6: Multiple Regression
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Multiple Regression MARE 250 Dr. Jason Turner.
Linear Regression MARE 250 Dr. Jason Turner.
Predictive Analysis in Marketing Research
REGRESSION AND CORRELATION
MARE 250 Dr. Jason Turner Correlation & Linear Regression.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chapter 15: Model Building
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Correlation and Regression Analysis
Simple Linear Regression Analysis
Relationships Among Variables
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Model selection Stepwise regression. Statement of problem A common problem is that there is a large set of candidate predictor variables. (Note: The examples.
M23- Residuals & Minitab 1  Department of ISM, University of Alabama, ResidualsResiduals A continuation of regression analysis.
Variable selection and model building Part II. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Chapter 14 Multiple Regression Models. 2  A general additive multiple regression model, which relates a dependent variable y to k predictor variables.
Shonda Kuiper Grinnell College. Statistical techniques taught in introductory statistics courses typically have one response variable and one explanatory.
Summarizing Bivariate Data
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Chapter 7 Relationships Among Variables What Correlational Research Investigates Understanding the Nature of Correlation Positive Correlation Negative.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Created by Erin Hodgess, Houston, Texas
Aron, Aron, & Coups, Statistics for the Behavioral and Social Sciences: A Brief Course (3e), © 2005 Prentice Hall Chapter 12 Making Sense of Advanced Statistical.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
MARE 250 Dr. Jason Turner Linear Regression. Linear regression investigates and models the linear relationship between a response (Y) and predictor(s)
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
1 Experimental Statistics - week 12 Chapter 12: Multiple Regression Chapter 13: Variable Selection Model Checking.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Variable Selection 1 Chapter 8 Variable Selection Terry Dielman Applied Regression Analysis:
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 Building the Regression Model –I Selection and Validation KNN Ch. 9 (pp )
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Variable selection and model building Part I. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Applied Quantitative Analysis and Practices LECTURE#28 By Dr. Osman Sadiq Paracha.
Venn diagram shows (R 2 ) the amount of variance in Y that is explained by X. Unexplained Variance in Y. (1-R 2 ) =.36, 36% R 2 =.64 (64%)
Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.
Model selection and model building. Model selection Selection of predictor variables.
MGS4020_Minitab.ppt/Jul 14, 2011/Page 1 Georgia State University - Confidential MGS 4020 Business Intelligence Regression Analysis By Using Minitab Jul.
Stats Methods at IC Lecture 3: Regression.
Multiple Regression.
Theme 6. Linear regression
Chapter 9 Multiple Linear Regression
Kin 304 Regression Linear Regression Least Sum of Squares
Multiple Regression.
BPK 304W Regression Linear Regression Least Sum of Squares
Making Sense of Advanced Statistical Procedures in Research Articles
CHAPTER 29: Multiple Regression*
Multiple Regression.
M248: Analyzing data Block D UNIT D2 Regression.
Chapter 11 Variable Selection Procedures
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

MARE 250 Dr. Jason Turner Multiple Regression

y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept b 1 = slope x = independent variable Urchin density = b 0 + b 1 (salinity)

Multiple regression allows us to learn more about the relationship between several independent or predictor variables and a dependent or criterion variable For example, we might be looking for a reliable way to estimate the age of AHI at the dock instead of waiting for laboratory analyses Multiple Regression y = b 0 + b 1 x y = b 0 + b 1 x 1 + b 2 x 2 …b n x n

In the social and natural sciences multiple regression procedures are very widely used in research Multiple regression allows the researcher to ask “what is the best predictor of...?” For example, researchers might want to learn what abiotic variables (temp, sal, DO, turb) are the best predictors of plankton abundance/diversity in Hilo Bay Or Which morphometric measurements are the best predictors of fish age Multiple Regression

The general computational problem that needs to be solved in multiple regression analysis is to fit a straight line to a number of points Multiple Regression In the simplest case - one dependent and one independent variable This can be visualized in a scatterplot

A line in a two dimensional or two-variable space is defined by the equation Y=a+b*X The Regression Equation In the multivariate case, when there is more than one independent variable, the regression line cannot be visualized in the two dimensional space, but can be computed rather easily

The smaller the variability of the residual values around the regression line relative to the overall variability, the better is our prediction Coefficient of determination (r 2 ) - If we have an R-square of 0.4 we have explained 40% of the original variability, and are left with 60% residual variability. Ideally, we would like to explain most if not all of the original variability Therefore - r 2 value is an indicator of how well the model fits the data (e.g., an r 2 close to 1.0 indicates that we have accounted for almost all of the variability with the variables specified in the model Residual Variance and R-square

Building Models via Stepwise Regression Stepwise model-building techniques for regression The basic procedures involve: (1) identifying an initial model (2) iteratively "stepping," that is, repeatedly altering the model at the previous step by adding or removing a predictor variable in accordance with the "stepping criteria," (3) terminating the search when stepping is no longer possible given the stepping criteria Stepwise Regression: When is too much – too much

For Example… We are interested in predicting values for Y based upon several X’s…Age of AHI based upon SL, BM, OP, PF We run multiple regression and get the equation: Age = SL BM OP PF We then run a STEPWISE regression to determine the best subset of these variables

S B O P Vars R-Sq R-Sq(adj) C-p S L M P F X X X X X X X X X X X X X X X X Response is Age How does it work… 1. Simplest model with the highest R 2 wins! 2. Use Mallows’ Cp to break the tie Who decides – YOU!

S B O P Vars R-Sq R-Sq(adj) C-p S L M P F X X X X X X X X X X X X X X X X Response is Age How does it work… You should also look a

How does it work… Stepwise Regression: Age versus SL, BM, OP, PF Alpha-to-Enter: 0.15 Alpha-to-Remove: 0.15 Response is Age on 4 predictors, with N = 84 Step Constant BM T-Value P-Value OP T-Value P-Value SL T-Value 1.96 P-Value S R-Sq R-Sq(adj) Mallows C-p

Who Cares? Stepwise analysis allows you (i.e. – computer) to determine which predictor variables (or combination of) best explain (can be used to predict) Y Much more important as number of predictor variables increase Helps to make better sense of complicated multivariate data