An introduction to basic multilevel modeling

Slides:



Advertisements
Similar presentations
Questions From Yesterday
Advertisements

Introduction Describe what panel data is and the reasons for using it in this format Assess the importance of fixed and random effects Examine the Hausman.
Hierarchical Linear Modeling: An Introduction & Applications in Organizational Research Michael C. Rodriguez.
Managerial Economics in a Global Economy
CORRELATION. Overview of Correlation u What is a Correlation? u Correlation Coefficients u Coefficient of Determination u Test for Significance u Correlation.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
The Simple Linear Regression Model: Specification and Estimation
Chapter 10 Simple Regression.
Chapter Topics Types of Regression Models
Chapter 11 Multiple Regression.
Topic 3: Regression.
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
Correlation and Regression Analysis
Simple Linear Regression and Correlation
Linear Regression 2 Sociology 5811 Lecture 21 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Introduction to Regression Analysis, Chapter 13,
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Analysis of Clustered and Longitudinal Data
3nd meeting: Multilevel modeling: introducing level 1 (individual) and level 2 (contextual) variables + interactions Subjects for today:  Intra Class.
Introduction to Multilevel Modeling Using SPSS
Regression and Correlation Methods Judy Zhong Ph.D.
Analysis of Variance. ANOVA Probably the most popular analysis in psychology Why? Ease of implementation Allows for analysis of several groups at once.
Advanced Business Research Method Intructor : Prof. Feng-Hui Huang Agung D. Buchdadi DA21G201.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Introduction Multilevel Analysis
CHAPTER 14 MULTIPLE REGRESSION
Introduction to Linear Regression
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Introduction to Multilevel Modeling Stephen R. Porter Associate Professor Dept. of Educational Leadership and Policy Studies Iowa State University Lagomarcino.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Review of Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Scatterplots & Regression Week 3 Lecture MG461 Dr. Meredith Rolfe.
University of Warwick, Department of Sociology, 2012/13 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 5 Multiple Regression.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
Lecture 10: Correlation and Regression Model.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
Methods of Presenting and Interpreting Information Class 9.
Predicting Energy Consumption in Buildings using Multiple Linear Regression Introduction Linear regression is used to model energy consumption in buildings.
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard)   Week 5 Multiple Regression  
Chapter 15 Panel Data Models.
The Correlation Coefficient (r)
12 Inferential Analysis.
I271B Quantitative Methods
Correlation and Regression
G Lecture 6 Multilevel Notation; Level 1 and Level 2 Equations
CHAPTER 29: Multiple Regression*
CHAPTER 26: Inference for Regression
6-1 Introduction To Empirical Models
Prepared by Lee Revere and John Large
From GLM to HLM Working with Continuous Outcomes
Multiple Regression Chapter 14.
12 Inferential Analysis.
CHAPTER 14 MULTIPLE REGRESSION
Product moment correlation
BEC 30325: MANAGERIAL ECONOMICS
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Analysing Means I: (Extending) Analysis.
3.2. SIMPLE LINEAR REGRESSION
BEC 30325: MANAGERIAL ECONOMICS
MGS 3100 Business Analysis Regression Feb 18, 2016
The Correlation Coefficient (r)
Presentation transcript:

An introduction to basic multilevel modeling

Two introductory points: 1. Hierarchical cross-national data structures are common within the field of sociology and political sciences, and ESS is a good example of such structures. Hierarchical cross-national data structures mean that we have information on the level of individuals and on the level of countries. The problem is how to explain one individual-level characteristic (variable) using both (a) other individual-level characteristics (variables), and (b) country-level characteristics (variables).

. 2. The old solution is to pool data from all countries and use country-level characteristics assigned to individuals as all other individual characteristics. There are three consequences of this (cross-sectional pooling) strategy: (a) For effects of individual-level variables we increase N so that these effects become significant while in the reality they might be not significant. (b) For effects of country-level variables we also increase N so that these effects become significant while in the reality they might be not significant. [Both (a) and (b) lead to increasing the probability of “false positive” = Error Type I, the error of rejecting a "correct" null hypothesis]. (c) Case dependency. Units of analysis are not independent – what is required by regular regression.

. Ignoring the statistical significance, for effects of individual-level variables, a pooling strategy can only be successful if countries are homogeneous. The extent that this condition really is met is indicated by the intra-class correlation. Usually this condition is poorly met. .

The basic two-level regression model with a continuous outcome An example taken from Stefan Dahlberg, Department of Political Science, Goteborg University (2007) We have opinion survey data from J countries with a number of respondents, i.e. voters i, ni, in each country j. On the individual level we have an outcome variable (Y) measured as the absolute distance between a voters self-placement on a left-right scale and the position, as ascribed by the voter, for the party voted for on the same scale. This dependent variable is called “proximity.” .

Two-levels We have two independent variables: X - education, and Z - the degree of proportionality of the electoral system. We assume that more educated people tend to vote consistently (high proximity). More proportional systems are expected to induce proximity voting since voters tend to view the elections as an expression of preferences rather than a process of selecting government. We have three variables on two different levels: Y proximity, X education (individual level) Z proportionality (country-level). .

Eq. 1 The simplest way to analyze the data is to make separate regressions in each country such as: (1.1.) Yij = α0j + β1jXij + eij The difference compared to a more ordinary regression model is that we assume that intercepts and slopes are varying between electoral systems, i.e. each country has its own intercept and slope, denoted by α0j and β1j. .

Interpretation The error term, eij is expected to have a mean of zero and a variance to be estimated. Since the intercept and slope coefficients are assumed to vary across countries they are often referred to as random coefficients. Across all countries the regression coefficients βj have a distribution with some mean and variance. .

Eq. 2-3 The next step in the hierarchical regression analysis is to introduce the country-level explanatory variable proportionality in an attempt to explain the variation of the regression coefficients α0j and β1j. (1.2) α0j = γ00 + γ01Zj + u0j and (1.3) β1j = γ10 + γ11Zj + u1j .

Interpretation Equation 1.2 predicts the average degree of proximity voting in a country (the intercept α0j ) by the degree of proportionality in the electoral system (Z). Equation 1.3 denotes that the relationship, as expressed by the coefficient β1j between proximity voting (Y) and level of education (X), is depending on the degree of proportionality within the electoral system (Z). The degree of proportionality is here functioning as an interactive variable where the relationship between proximity voting and education varies according to the values on the second level variable proportionality.

. The u-terms u0j and u1j in equation 1.2 and 1.3 are random residual error terms at the country level. The residual errors uj are assumed to have a mean of zero and to be independent from the residual errors eij at the individual level. The variance of the residual errors u0j is specified as σ2u0 and the variance of the residual errors u1j is specified as σ2u1.

. In equation 1.2 and 1.3 the regression coefficients γ are not assumed to vary across countries since they have the denotation j in order to indicate to which country they belong. As they apply to all countries the coefficients are referred to as fixed effects. All the between-country variation left in the β coefficients, after predicting these with the country variable proportionality (Zj), is then assumed to be the residual error variation. This is captured by the error term uj, which are denoted with j to indicate which country it belongs to. .

Eq. 4 The equations 1.2 and 1.3 can then be substituted into equation 1.1 and by rearranging them we have: (1.4) Yij = γ00 + γ10Xij + γ01Zj + γ11XijZj + u1jXij + u0j + eij .

+ γ11 educationij * proportionalityj + u1j educationij + u0j + eij Interpretation In terms of variable labels the equation 1.4 states: proximityij = γ00 + γ10 educationij + γ01 proportionalityj + γ11 educationij * proportionalityj + u1j educationij + u0j + eij

. The first part of equation 1.4: γ00 + γ10Xij + γ01Zj + γ11XijZj is the fixed coefficients/effects. The term XijZj is an interaction term that is included in the model as a consequence of the modeling of the varying regression slope β1j on individual level variable Xij with the country level Zj.

. The last part of the equation 1.4 u1jXij + u0j + eij is a part of the random coefficients. The fact that the explanatory variable Xij is multiplied with the error term, u1jXij, the resulting total error will be different for different values of Xij.

. The error terms are heteroscedastic instead of homoscedastic as assumed in ordinary regression models where the residual errors are assumed to be independent of the values of the explanatory variable. [Random variables are heteroscedastic if they have different variances for the relevant subgrups. The complementary concept is called homoscedasticity. Note: The alternative spelling homo- or heteroskedasticity is equally correct and is also used frequently. The term means "differing variance" and comes from the Greek "hetero" (different) and "skedasis" (dispersion).]

. Dealing with the problem of heteroscedasticy is one of the main reasons for why multilevel models are preferable over regular OLS models when analyzing hierarchical nested data used.

Known in literature under a variety of names SUMMING UP: Known in literature under a variety of names  Hierarchical linear model (HLM)  Random coefficient model  Variance component model  Multilevel model  Contextual analysis  Mixed linear model  Mixed effects model Acknowldegements: Materials in the following are based on Prof. Joop J. Hox introductory lectures on HLM delivered at the 2011 QMSS2 Summer School in Leuven 19

Multi-level data structure Groups at different levels may have different sizes Response (outcome) variable at lowest level Explanatory variables at all levels The statistical model assumes sampling at all levels Ex: education family longitudinal level 3 schools classes level 2 classes families level 1 pupils members occasions (waves)

Problems with Hierarchical Data Structure I Assume DV (response) variable on lowest level (level 1). We want to predict this, using explanatory variables at all available levels - How? - What is the proper sample size? - What if relations are not the same in different groups? Traditional approaches: Disaggregate all variables to the lowest level Do standard analyses (anova, multiple regression) Aggregate all variables to the highest level Ancova with groups as factor Some improvements: explanatory variables as deviations from their group mean have both deviation score & disaggregated group mean as predictor (separates individual and group effects) What is wrong with this? What if relations are not the same in different groups? Ex: relationship SIĘ – educ may be different in differnet schools 21

Problems with Hierarchical Data Structure II Multiple Regression assumes:  independent observations  independent error terms  equal variances of errors for all observations (assumption of homoscedastic errors)  normal distribution for errors With hierarchical data  observations are not independent  errors are not independent  different observations may have errors with different variances (heteroscedastic errors) You violate the first 2 22

Problems with Hierarchical Data Structure III Observations in the same group are generally not independent - they tend to be more similar than observations from different groups - selection, shared history, contextual group effects The degree of similarity is indicated by the intraclass correlation rho: r Kish: roh (rate of homogeneity) Standard statistical tests are not at all robust against violation of the independence assumption That’s why we need special multilevel techniques.

Graphical Pictureof 2-level regression Model School level Student level Outcome variable on pupil level - Explanatory variables at both levels: individual & group - Residual error at individual level - Plus residual error at school level School size error error Grade Student gender The relation btw gender and grade may be different across schools; error at school level not accounted for in standard regression

Graphical Pictureof 2-level regression Model School level Student level Essential points: Explanatory variables characterize individuals and/or groups Average value of individual variables may differ across groups = most variables have both within-group and between-group variation Groups do make a difference: the effect of an individual explanatory variable may be different in different groups School size error error Grade Student gender The relation btw gender and grade may be different across schools; error at school level not accounted for in standard regression

Yij= [g00+ g10Xij + g01Zj + g11ZjXij] + [u1jXij+ u0j+ eij] Assumptions Yij= [g00+ g10Xij + g01Zj + g11ZjXij] + [u1jXij+ u0j+ eij]  Individual level errors eij independent; normal distribution with mean zero and same variance se² in all groups  Group level errors u.j independent; multivariate normal distribution with means zero and (co)variances su² in covariance matrix W (omega)  Group level errors u.j independent from individual errors eij - plus usual assumptions of multiple regression analysis - linear relations, explanatory variables measured without error

Types of HLM Models Intercept-only Model Intercept only model (null model, baseline model) Contains only intercept and corresponding error terms At the lowest (individual) level we have Yij= b0j+ eij and at the second level b0j= g00+ u0j hence Yij= g00+ u0j+ eij Used to decompose the total variance and to compute the intraclass correlation r (rho) r = (group level variance / total variance) r = expected correlation between 2 individuals in same group

Only fixed effects for level-1 explanatory variables: Fixed Model: Only fixed effects for level-1 explanatory variables: slopes are assumed not to vary across groups At the lowest (individual) level we have Yij = b0j+ b1jXij+ eij and at the second level b0j = g00+ u0j and b1j= g10  hence Yij= g00+ g10Xij+ u0j+ eij Intercepts vary across groups, slopes are the same Similar to ANCOVA with random grouping factor also called variance component model  Fixed effects for level-2 explanatory variables added next 28

Random Coefficient Model Assumes intercept and slopes vary across groups Yij= b0j+ b1jXij+ eij and at the second level b0j= g00+ u0j and b1j= g10+ u1j hence Yij= g00+ g10Xij+ u1jXij+ u0j+ eij The full multilevel model adds interactions to model the slope variation 29

Full Multilevel Regression Model Explanatory variables at all levels Higher level variables predict variation of lowest level intercept and slopes At the lowest (individual) level we have Yij= b0j+ b1jXij+ eij  and at the second level b0j= g00+ g01Zj+ u0j and b1j= g10+ g11Zj+ u1j hence Yij= g00+ g10Xij+ g01Zj+ g11ZjXij+ u1jXij+ u0j+ eij Predicting the intercept implies a direct effect Predicting slopes implies cross-level interactions 30