3nd meeting: Multilevel modeling: introducing level 1 (individual) and level 2 (contextual) variables + interactions Subjects for today:  Intra Class.

Slides:



Advertisements
Similar presentations
To go further: intra- versus interindividual variability.
Advertisements

Hierarchical Linear Modeling: An Introduction & Applications in Organizational Research Michael C. Rodriguez.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
AMMBR from xtreg to xtmixed (+checking for normality, random slopes)
5nd meeting: Multilevel modeling: Summary & Extra’s Subjects for today:  How to do multilevel analysis: a 5-step-approach  Interaction, cross-level interactions,
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Linear Regression t-Tests Cardiovascular fitness among skiers.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Review of the Basic Logic of NHST Significance tests are used to accept or reject the null hypothesis. This is done by studying the sampling distribution.
Multiple Regression Fenster Today we start on the last part of the course: multivariate analysis. Up to now we have been concerned with testing the significance.
Objectives (BPS chapter 24)
Advanced Methods and Models in Behavioral Research – 2014 Been there / done that: Stata Logistic regression (……) Conjoint analysis Coming up: Multi-level.
July 1, 2008Lecture 17 - Regression Testing1 Testing Relationships between Variables Statistics Lecture 17.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Chapter 13 Multiple Regression
Chapter 10 Simple Regression.
Chapter 12 Multiple Regression
Chapter 4 Multiple Regression.
Multiple Regression Models
The Simple Regression Model
Intro to Statistics for the Behavioral Sciences PSYC 1900
Treatment Effects: What works for Whom? Spyros Konstantopoulos Michigan State University.
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
REGRESSION AND CORRELATION
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Simple Linear Regression Analysis
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Variance and covariance Sums of squares General linear models.
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Lecture 5 Correlation and Regression
Regression and Correlation Methods Judy Zhong Ph.D.
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
Inference for regression - Simple linear regression
Statistics Primer ORC Staff: Xin Xin (Cindy) Ryan Glaman Brett Kellerstedt 1.
Advanced Methods and Models in Behavioral Research – 2010/2011 AMMBR course design CONTENT METHOD Y is 0/1 conjoint analysis logistic regression multi-level.
Introduction Multilevel Analysis
Introduction to Linear Regression
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
1st meeting: Multilevel modeling: introduction Subjects for today:  Basic statistics (testing)  The difference between regression analysis and multilevel.
Testing Hypotheses about Differences among Several Means.
Multilevel Linear Modeling aka HLM. The Design We have data at two different levels In this case, 7,185 students (Level 1) Nested within 160 Schools (Level.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT OSMAN BIN SAIF Session 26.
2nd meeting: Multilevel modeling: intra class correlation Subjects for today:  Multilevel data base construction  The difference between single level.
HLM Models. General Analysis Strategy Baseline Model - No Predictors Model 1- Level 1 Predictors Model 2 – Level 2 Predictors of Group Mean Model 3 –
Data Analysis in Practice- Based Research Stephen Zyzanski, PhD Department of Family Medicine Case Western Reserve University School of Medicine October.
Introduction to the t Test Part 1: One-sample t test
Multiple Logistic Regression STAT E-150 Statistical Methods.
Advanced Methods and Models in Behavioral Research – 2009/2010 AMMBR course design CONTENT METHOD Y is 0/1 conjoint analysis logistic regression multi-level.
Outline of Today’s Discussion 1.The Distribution of Means (DOM) 2.Hypothesis Testing With The DOM 3.Estimation & Confidence Intervals 4.Confidence Intervals.
Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and Alison Kelly Copyright © 2014 by McGraw-Hill Higher Education. All rights.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Example x y We wish to check for a non zero correlation.
Outline of Today’s Discussion 1.Seeing the big picture in MR: Prediction 2.Starting SPSS on the Different Models: Stepwise versus Hierarchical 3.Interpreting.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Lecture 7: Bivariate Statistics. 2 Properties of Standard Deviation Variance is just the square of the S.D. If a constant is added to all scores, it has.
Multiple Regression Scott Hudson January 24, 2011.
Stats Methods at IC Lecture 3: Regression.
Lecture #25 Tuesday, November 15, 2016 Textbook: 14.1 and 14.3
Review Guess the correlation
Regression 11/6.
Regression 10/29.
Applied Biostatistics: Lecture 2
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

3nd meeting: Multilevel modeling: introducing level 1 (individual) and level 2 (contextual) variables + interactions Subjects for today:  Intra Class Correlation once more  Including individual (level 1) variables  Including contextual (level 2) variables  (cross level) interactions

Last meeting: Intra Class Correlation: a fraction, it tells us what the relative share of level 2 variance is in the total variance: ICC = level 2 / level 2 + level 1 variance ICC is always between 0 (only level 1 variance, no clustering) and 1 (only level 2 variance) We test with a Chi-square test: difference in -2 loglikelihood between model with both level 1 and level 2 variance and model with level 2 variance set to zero. 2

Multilevel null model: ICC = / ( ) =.23 3

Test whether ICC is significant or whether level 2 variance is significant different from zero we perform a Chi-square test: -2 loglikelihood from 1 level model - -2 loglikelihood from 2 level model: – = 132 with 1 df. Which is highly significant. This test is superior to Z-test in SPSS because the latter assumes normal distribution. Note: we test one sided because outcome is always zero or higher Model with one level only: 4

Testing in MLwin with Chi-square (more info in document about testing, see “statistical testing in Mlwin.pdf” Type cprob 10 1 and press [Enter] Note that must be divided by 2 voor level 2 variance testing 5

So we have taken one step in the whole procedure: we tested whether there is significant level 2 variance. In the example with schools we indeed have a level 2 variance. The next step is to add level 1 variables to the equation: Y = b0j + b1 * level 1 variable + eij + uoj Why? First, we might have hypotheses about level 1 variables, for instance we assume that white students perform best on a math test compared to other students. Second, we might expect that level 1 variables can explain level 2 variance. How is that? 6

Suppose So we have had one step in the whole procedure: we tested whether there is level 2 variance. In the example with schools we indeed have a level 2 variance The next step is to add level 1 variables to the equation: Y = b0 + b1 * level 1 variable + eij + uoj Why? First, we might have hypotheses about level 1 variables, for instance we assume that white students perform better on a math test. Second, we might expect that level 1 variables can explain level 2 variance. How is that? 7

Well, suppose being white indeed increases the chances to have a higher grade at a math test. Let us further assume that the fraction whites across schools varies. So assume school 1 has 80% whites and school 2 has 40% whites and we have a positive effect for being white on a math test. This must result in a lower average grade for school 2 compared to school 1, all other things being equal. This is called a compositional effect: the mere fact that whites perform better on a math test, school 1 (with 80% whites) will perform better than school 2 (with 40% whites). This eventually means that part of the difference between the school’s average is a result of the race composition. As a consequence introducing the individual level variable ‘race’ will REDUCE level 2 variance!! 8

Including the dummy ‘white’ reduces level 2 variance while it was in the null model Level 1 variance is also reduced (was ) Note that -2 loglikelihood dropped to was Being white increases Grade by 4.7! 9

Testing the effect of being white: 2 options: One is traditional t-test: b / se = t with df = n – p (where n=sample, p=number of level 1 predictors: / = with 519 – 1 = 518. Data manipulation  Command interface and type: tprob Second option: take the -2 loglikelihood difference from model with and without level 1 predictor and use Chi-square test: = with df = 1 (in linear models it is equal to t-test because square root taken from = 3.946) So type cprob Take half of that because directional hypothesis! 10

Of course one will introduce more level 1 variables. First, to test hypotheses, second to find a (partly) explanation for level 2 variance (we now use the full dataset: LOGSCHOOL.SAV: 415 schools, >8000 pupils. After we included all relevant level 1 predictors we move on to step number 2: introduce level 2 variables: here the % whites (variable name: race2_mean) 415 schools, t-test: 7.555/.8 = 9!! With df = n2 – 1 – 1=

You see there is a difference between a compositional effect from being white and the level 2 effect of %whites. The first effect is just due to individual behavior that taken together influences the school’s average on a math test. The second effect tells us that individuals perform better when a school has a high percentage of whites. Now let us finally include public versus private school (students in public schools do worse: estimate ). 415 schools, df=415 –

Now we have done 3 things: 1) estimate ICC, 2) included level 1 variables, 3) include level 2 variables. But now…. We might think that the effect of homework varies across schools: 13

Let us get back to the model with no random effect of homework. So effect of homework is constant across schools: 14

There are two extra parameters: 1) the variance of the slope for Homework (not all with the same slope angle, even -! 2) a co-variance between slope and intercept (high on intercept, lower slope) Effect of homework set random across schools: 15

Ok, so we know that the effect of homework depends on what school you are in: in some schools the effect is low, while in others it is high. WHY is that? One explanation: it is so because in private schools quality of staff is higher so making homework is less important. This would imply that the effect of homework is stronger when school is public: Effect of homework when school = private equals.914. When public the effect doubles = = 1.831!! 16