Wednesday PM  Presentation of AM results  Multiple linear regression Simultaneous Simultaneous Stepwise Stepwise Hierarchical Hierarchical  Logistic.

Slides:



Advertisements
Similar presentations
The SPSS Sample Problem
Advertisements

Regression analysis Linear regression Logistic regression.
Regression With Categorical Variables. Overview Regression with Categorical Predictors Logistic Regression.
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
Logistic Regression Part I - Introduction. Logistic Regression Regression where the response variable is dichotomous (not continuous) Examples –effect.
MULTIPLE REGRESSION. OVERVIEW What Makes it Multiple? What Makes it Multiple? Additional Assumptions Additional Assumptions Methods of Entering Variables.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
Discrim Continued Psy 524 Andrew Ainsworth. Types of Discriminant Function Analysis They are the same as the types of multiple regression Direct Discrim.
Multiple Regression.
LECTURE 5 MULTIPLE REGRESSION TOPICS –SQUARED MULTIPLE CORRELATION –B AND BETA WEIGHTS –HIERARCHICAL REGRESSION MODELS –SETS OF INDEPENDENT VARIABLES –SIGNIFICANCE.
Biol 500: basic statistics
Handling Categorical Data. Learning Outcomes At the end of this session and with additional reading you will be able to: – Understand when and how to.
An Introduction to Logistic Regression
Chapter 15: Model Building
Lecture 5: Simple Linear Regression
Multiple Regression – Basic Relationships
Logistic Regression Chapter 8.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Logistic Regression – Basic Relationships
Multiple Regression Dr. Andy Field.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Split Sample Validation General criteria for split sample validation Sample problems.
STAT E-150 Statistical Methods
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
Categorical Data Prof. Andy Field.
Selecting the Correct Statistical Test
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
LEARNING PROGRAMME Hypothesis testing Intermediate Training in Quantitative Analysis Bangkok November 2007.
Hierarchical Binary Logistic Regression
Multinomial Logistic Regression Basic Relationships
Discriminant Analysis
Logistic (regression) single and multiple. Overview  Defined: A model for predicting one variable from other variable(s).  Variables:IV(s) is continuous/categorical,
Review of Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
PS 225 Lecture 20 Linear Regression Equation and Prediction.
Aron, Aron, & Coups, Statistics for the Behavioral and Social Sciences: A Brief Course (3e), © 2005 Prentice Hall Chapter 12 Making Sense of Advanced Statistical.
Review for Final Examination COMM 550X, May 12, 11 am- 1pm Final Examination.
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
LOGISTIC REGRESSION Binary dependent variable (pass-fail) Odds ratio: p/(1-p) eg. 1/9 means 1 time in 10 pass, 9 times fail Log-odds ratio: y = ln[p/(1-p)]
Logistic Regression Analysis Gerrit Rooks
Dates Presentations Wed / Fri Ex. 4, logistic regression, Monday Dec 7 th Final Tues. Dec 8 th, 3:30.
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Applied Quantitative Analysis and Practices LECTURE#28 By Dr. Osman Sadiq Paracha.
Nonparametric Statistics
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 18 Multivariate Statistics.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
STATISTICAL TESTS USING SPSS Dimitrios Tselios/ Example tests “Discovering statistics using SPSS”, Andy Field.
Multiple Regression Scott Hudson January 24, 2011.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Chapter 13 LOGISTIC REGRESSION. Set of independent variables Categorical outcome measure, generally dichotomous.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Nonparametric Statistics
Regression Analysis.
BINARY LOGISTIC REGRESSION
Logistic Regression APKC – STATS AFAC (2016).
Correlation, Bivariate Regression, and Multiple Regression
Regression Analysis.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Analysis of Covariance (ANCOVA)
Regression.
Making Sense of Advanced Statistical Procedures in Research Articles
بحث في التحليل الاحصائي SPSS بعنوان :
Multiple Regression – Part II
Multiple logistic regression
Nonparametric Statistics
Introduction to Logistic Regression
Multiple Regression – Split Sample Validation
Regression Analysis.
Presentation transcript:

Wednesday PM  Presentation of AM results  Multiple linear regression Simultaneous Simultaneous Stepwise Stepwise Hierarchical Hierarchical  Logistic regression

Multiple regression  Multiple regression extends simple linear regression to consider the effects of multiple independent variables (controlling for each other) on the dependent variable.  The line fit is: Y = b 0 + b 1 X 1 + b 2 X 2 + b 3 X 3 + …  The coefficients (b i ) tell you the independent effect of a change in one dependent variable on the independent variable, in natural units.

Multiple regression in SPSS  Same as simple linear regression, but put more than one variable into the independent box.  Equation output has a line for each variable: Coefficients: Predicting Q2 from Q3, Q4, Q5 UnstandardizedStandardized UnstandardizedStandardized BSEBetatSig. (Constant) Q Q Q Q Q Q  Unstandardized coefficients are the average effect of each independent variable, controlling for all other variables, on the dependent variable.

Standardized coefficients  Standardized coefficients can be used to compare effect sizes of the independent variables within the regression analysis.  In the preceding analysis, a change of 1 standard deviation in Q3 has over 6 times the effect of a change of 1 sd in Q5 and over 30 times the effect of a change of 1 sd in Q4.  However,  s are not stable across analyses and can’t be compared.

Stepwise regression  In simultaneous regression, all independent variables are entered in the regression equation.  In stepwise regression, an algorithm decides which variables to include.  The goal of stepwise regression is to develop the model that does the best prediction with the fewest variables.  Ideal for creating scoring rules, but atheoretical and can capitalize on chance (post-hoc modeling)

Stepwise algorithms  In forward stepwise regression, the equation starts with no variables, and the variable that accounts for the most variance is added first. Then the next variable that can add new variance is added, if it adds a significant amount of variance, etc.  In backward stepwise regression, the equation starts with all variables; variables that don’t add significant variance are removed.  There are also hybrid algorithms that both add and remove.

Stepwise regression in SPSS  Analyze…Regression…Linear  Enter dependent variable and independent variables in the independents box, as before  Change “Method” in the independents box from “Enter” to: Forward Forward Backward Backward Stepwise Stepwise

Hierarchical regression  In hierarchical regression, we fit a hierarchy of regression models, adding variables according to theory and checking to see if they contribute additional variance.  You control the order in which variables are added  Used for analyzing the effect of dependent variables on independent variables in the presence of moderating variables.  Also called path analysis, and equivalent to analysis of covariance (ANCOVA).

Hierarchical regression in SPSS  Analyze…Regression…Linear  Enter dependent variable, and the independent variables you want added for the smallest model  Click “Next” in the independents box  Enter additional independent variables  …repeat as required…

Hierarchical regression example  In the hyp data, there is a correlation of between case-based course and final exam.  Is the relationship between final exam score and course format moderated by midterm exam score? FinalCase-based Midterm

Hierarchical regression example  To answer the question, we: Predict final exam from midterm and format (gives us the effect of format, controlling for midterm, and the effect of midterm, controlling for format) Predict final exam from midterm and format (gives us the effect of format, controlling for midterm, and the effect of midterm, controlling for format) Predict midterm from format (gives us the effect of format on midterm) Predict midterm from format (gives us the effect of format on midterm)  After running each regression, write the  s on the path diagram:

Predict final from midterm, format Coefficients BSEBeta tSig. BSEBeta tSig. (Constant) Case-based course midterm exam score FinalCase-based Midterm *.207 *

Coefficients BSEBeta tSig. BSEBeta tSig. (Constant) Case-based course  Conclusions: The course format affects the final exam both directly and through an effect on the midterm exam. In both cases, lecture courses yielded higher scores. Predict midterm from format FinalCase-based Midterm *.207 * *

 Linear regression fits a line.  Logistic regression fits a cumulative logistic function S-shaped S-shaped Bounded by [0,1] Bounded by [0,1]  This function provides a better fit to binomial dependent variables (e.g. pass/fail)  Predicted dependent variable represents the probability of one category (e.g. pass) based on the values of the independent variables. Logistic regression

Logistic regression in SPSS  Analyze…Regression…Binary logistic (or multinomial logistic)  Enter dependent variable and independent variables  Output will include: Goodness of model fit (tests of misfit) Goodness of model fit (tests of misfit) Classification table Classification table Estimates for effects of independent variables Estimates for effects of independent variables  Example: Voting for Clinton vs. Bush in 1992 US election, based on sex, age, college graduate

Logistic regression output   Goodness of fit measures: -2 Log Likelihood (lower is better) Goodness of Fit (lower is better) Cox & Snell - R^2.012(higher is better) Nagelkerke - R^2.016(higher is better) Chi-Square df Significance Model (A significant chi-square indicates poor fit (significant difference between predicted and observed data), but most models on large data sets will have significant chi-square)

Logistic regression output Classification Table The Cut Value is.50 Predicted Bush Clinton Percent Correct B | C Observed Bush B | 0 | 661 |.00% Clinton C | 0 | 907 | % Overall 57.84%

Logistic regression output VariableBS.E.WalddfSigRExp(B) FEMALE OVER COLLGRAD Constant  B is the coefficient in log-odds; Exp(B) = e B gives the effect size as an odds ratio.  Your odds of voting for Clinton are 1.54 times greater if you’re a woman than a man.

Wednesday PM assignment  Using the semantic data set: Perform a regression to predict total score from semantic classification. Interpret the results. Perform a regression to predict total score from semantic classification. Interpret the results. Perform a one-way ANOVA to predict total score from semantic classification. Are the results different? Perform a one-way ANOVA to predict total score from semantic classification. Are the results different? Perform a stepwise regression to predict total score. Include semantic classification, number of distinct semantic qualifiers, reasoning, and knowledge. Perform a stepwise regression to predict total score. Include semantic classification, number of distinct semantic qualifiers, reasoning, and knowledge. Perform a logistic regression to predict correct diagnosis from total score and number of distinct semantic qualifiers. Interpret the results. Perform a logistic regression to predict correct diagnosis from total score and number of distinct semantic qualifiers. Interpret the results.