Mixture Modeling Chongming Yang Research Support Center FHSS College.

Slides:



Advertisements
Similar presentations
Structural Equation Modeling. What is SEM Swiss Army Knife of Statistics Can replicate virtually any model from “canned” stats packages (some limitations.
Advertisements

Structural Equation Modeling Using Mplus Chongming Yang Research Support Center FHSS College.
Latent Growth Modeling Chongming Yang Research Support Center FHSS College.
Mixture modelling of continuous variables. Mixture modelling So far we have dealt with mixture modelling for a selection of binary or ordinal variables.
Outline input analysis input analyzer of ARENA parameter estimation
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
Visual Recognition Tutorial
Chapter 10 Simple Regression.
Visual Recognition Tutorial
Structural Equation Modeling Intro to SEM Psy 524 Ainsworth.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Introduction to Multilevel Modeling Using SPSS
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences.
EM and expected complete log-likelihood Mixture of Experts
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
1 1 Slide © 2012 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
A statistical model Μ is a set of distributions (or regression functions), e.g., all uni-modal, smooth distributions. Μ is called a parametric model if.
1 1 Slide © 2004 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Social patterning in bed-sharing behaviour A longitudinal latent class analysis (LLCA)
Learning Theory Reza Shadmehr logistic regression, iterative re-weighted least squares.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Multilevel Linear Models Field, Chapter 19. Why use multilevel models? Meeting the assumptions of the linear model – Homogeneity of regression coefficients.
Latent Growth Curve Modeling In Mplus: An Introduction and Practice Examples Part II Edward D. Barker, Ph.D. Social, Genetic, and Developmental Psychiatry.
Logistic Regression Database Marketing Instructor: N. Kumar.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Maximum Likelihood Estimation Methods of Economic Investigation Lecture 17.
Univariate Linear Regression Problem Model: Y=  0 +  1 X+  Test: H 0 : β 1 =0. Alternative: H 1 : β 1 >0. The distribution of Y is normal under both.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Roghayeh parsaee  These approaches assume that the study sample arises from a homogeneous population  focus is on relationships among variables 
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Lecture 2: Statistical learning primer for biologists
ECE 8443 – Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem Proof EM Example – Missing Data Intro to Hidden Markov Models.
Latent regression models. Where does the probability come from? Why isn’t the model deterministic. Each item tests something unique – We are interested.
University Rennes 2, CRPCC, EA 1285
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Logistic regression. Recall the simple linear regression model: y =  0 +  1 x +  where we are trying to predict a continuous dependent variable y from.
Tutorial I: Missing Value Analysis
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Chapter 17 STRUCTURAL EQUATION MODELING. Structural Equation Modeling (SEM)  Relatively new statistical technique used to test theoretical or causal.
The SweSAT Vocabulary (word): understanding of words and concepts. Data Sufficiency (ds): numerical reasoning ability. Reading Comprehension (read): Swedish.
BINARY LOGISTIC REGRESSION
Structural Equation Modeling using MPlus
Probability Theory and Parameter Estimation I
Basic Estimation Techniques
Classification of unlabeled data:
LECTURE 10: EXPECTATION MAXIMIZATION (EM)
Data Mining Lecture 11.
Latent Variables, Mixture Models and EM
Slides by JOHN LOUCKS St. Edward’s University.
Basic Estimation Techniques
Bayesian Models in Machine Learning
Mathematical Foundations of BME Reza Shadmehr
EC 331 The Theory of and applications of Maximum Likelihood Method
Introduction to Logistic Regression
Latent Variable Mixture Growth Modeling in Mplus
Parametric Methods Berlin Chen, 2005 References:
Learning From Observed Data
Biointelligence Laboratory, Seoul National University
Applied Statistics and Probability for Engineers
Structural Equation Modeling
St. Edward’s University
Presentation transcript:

Mixture Modeling Chongming Yang Research Support Center FHSS College

Mixture of Distributions

Classification Techniques Latent Class Analysis (categorical indicators) Latent Profile Analysis (continuous Indicators) Finite Mixture Modeling (multivariate normal variables) …

Integrate Classification Models into Other Models Mixture Factor Analysis Mixture Regressions Mixture Structural Equation Modeling Growth Mixture Modeling Multilevel Mixture Modeling

Disadvantages of Multi-steps Practice Multistep practice – Run classification model – Save membership Variable – Model membership variable and other variables Disadvantages – Biases in parameter estimates – Biases in standard errors Significance Confidence Intervals

Latent Class Analysis (LCA) Setting – Latent trait assumed to be categorical – Trait measured with multiple categorical indicators – Example: drug addiction, Schizophrenia Aim – Identify heterogeneous classes/groups – Estimate class probabilities – Identify good indicators of classes – Relate covariates to Classes

Graphic LCA Model Categorical Indicators u: u 1, u 2,u 3, …u r Categorical Latent Variable C: C =1, 2, …, or K

Probabilistic Model Assumption: Conditional independence of u so that interdependence is explained by C like factor analysis model An item probability Joint Probability of all indicators

LCA Parameters Number of Classes -1 Item Probabilities -1

Class Means (Logit)

Latent Class Analysis with Covariates

Posterior Probability (membership/classification of cases)

Estimation Maximum Likelihood estimation via Expectation-Maximization algorithm – E (expectation) step: compute average posterior probabilities for each class and item – M (maximization) step: estimate class and item parameters – Iterate EM to maximize the likelihood of the parameters

Test against Data O = observed number of response patterns E = model estimated number of response patterns Pearson Chi-square based on likelihood ratio

Determine Number of Classes Substantive theory (parsimonious, interpretable) Predictive validity Auxiliary variables / covariates Statistical information and tests – Bayesian Information Criterion (BIC) – Entropy – Testing K against K-1 Classes Vuong-Lo-Mendell-Rubin likelihood-ratio test Bootstrapped likelihood ratio test

Bayesian Information Criterion (BIC) L = likelihood h = number of parameters N = sample size Choose model with smallest BIC BIC Difference > 4 appreciable

Quality of Classification

Testing K against K-1 Classes Bootstrapped likelihood ratio test LRT = 2[logL(model 1)- logL(model2)], where model 2 is nested in model 1. Bootstrap Steps: 1.Estimate LRT for both models 2.Use bootstrapped samples to obtain distributions for LRT of both models 3.Compare LRT and get p values

Testing K against K-1 Classes Vuong-Lo-Mendell-Rubin likelihood-ratio test

Determine Quality of Indicators Good indicators – Item response probability is close to 0 or 1 in each class Bad indicators – Item response probability is high in more than one classes, like cross-loading in factor analysis – Item response probability is low in all classes like low-loading in factor analysis

LCA Examples LCA LCA with covariates Class predicts a categorical outcome

Save Membership Variable Variable: idvar = id; Output: Savedata: File = cmmber.txt; Save = cprob;

Latent Profile Analysis

Finite Mixture Modeling ( multivariate normal variables) Finite = finite number of subgroups/classes Variables are normally distributed in each class Means differ across classes Variances are the same across Covariances can differ without restrictions or equal with restrictions across classes Latent profile can be special case with covariances fixed at zero.

Mixture Factor Analysis Allow one to examine measurement properties of items in heterogeneous subgroups / classes Measurement invariance is not required assuming heterogeneity Factor structure can change See Mplus outputs

Factor Mixture Analysis Parental Control Parental Acceptance Feel people in your family understand you Feel you want to leave home Feel you and your family have fun together Feel that your family pay attention to you Feel your parents care about you Feel close to your mother Feel close to your father Parents let you make your own decisions about the time you must be home on weekend nights Parents let you make your own decisions about the people you hang around with Parents let you make your own decisions about what you wear Parents let you make your own decisions about which television programs you watch Parents let you make your own decisions about what time you go to bed on week nights Parents let you make your own decisions about what you eat

Two dimensions of Parenting

Mixture SEM See mixture growth modeling

Mixture Modeling with Known Classes Identify hidden classes within known groups Under nonrandomized experiments – Impose equality constraints on covariates to identify similar classes from known groups – Compare classes that differ in covariates