ML estimation in Gaussian graphical models

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Factorial Mixture of Gaussians and the Marginal Independence Model Ricardo Silva Joint work-in-progress with Zoubin Ghahramani.
Copula Regression By Rahul A. Parsa Drake University &
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
2 – In previous chapters: – We could design an optimal classifier if we knew the prior probabilities P(wi) and the class- conditional probabilities P(x|wi)
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Object Orie’d Data Analysis, Last Time Finished NCI 60 Data Started detailed look at PCA Reviewed linear algebra Today: More linear algebra Multivariate.
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Basics of Statistical Estimation. Learning Probabilities: Classical Approach Simplest case: Flipping a thumbtack tails heads True probability  is unknown.
A gentle introduction to Gaussian distribution. Review Random variable Coin flip experiment X = 0X = 1 X: Random variable.
Different chi-squares Ulf H. Olsson Professor of Statistics.
Visual Recognition Tutorial
Linear and generalised linear models
Chapter 1 Systems of Linear Equations
July 3, Department of Computer and Information Science (IDA) Linköpings universitet, Sweden Minimal sufficient statistic.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
Lecture 7 1 Statistics Statistics: 1. Model 2. Estimation 3. Hypothesis test.
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
Linear and generalised linear models
10.1 Gaussian Elimination Method
Bayesian Wrap-Up (probably). Administrivia Office hours tomorrow on schedule Woo hoo! Office hours today deferred... [sigh] 4:30-5:15.
Appendix A of: Symmetry and Lattice Conditional Independence in a Multivariate Normal Distribution. by Andersson& Madsen. Presented by Shaun Deaton.
Stats & Linear Models.
Today Wrap up of probability Vectors, Matrices. Calculus
Maximum Likelihood Estimation
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
1 Reproducing Kernel Exponential Manifold: Estimation and Geometry Kenji Fukumizu Institute of Statistical Mathematics, ROIS Graduate University of Advanced.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
A Generalization of PCA to the Exponential Family Collins, Dasgupta and Schapire Presented by Guy Lebanon.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Chapter 1 Systems of Linear Equations Linear Algebra.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Discretization of Continuous-Time State Space Models
Conditional Expectation
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Commuting birth-and-death processes Caroline Uhler Department of Statistics UC Berkeley (joint work with Steven N. Evans and Bernd Sturmfels) MSRI Workshop.
Biointelligence Laboratory, Seoul National University
CS479/679 Pattern Recognition Dr. George Bebis
Chapter 3: Maximum-Likelihood Parameter Estimation
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
10.8 Systems of Second-Degree Equations and Inequalities
Markov Chains Mixing Times Lecture 5
Parameter Estimation 主講人:虞台文.
CH 5: Multivariate Methods
Menglong Li Ph.d of Industrial Engineering Dec 1st 2016
STATISTICAL INFERENCE PART I POINT ESTIMATION
Maximum Likelihood Estimation
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Course Outline MODEL INFORMATION COMPLETE INCOMPLETE
EC 331 The Theory of and applications of Maximum Likelihood Method
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
The Multivariate Normal Distribution, Part 2
Singular Value Decomposition SVD
Generally Discriminant Analysis
Computing and Statistical Data Analysis / Stat 7
Learning From Observed Data
12. Principles of Parameter Estimation
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Outline Texture modeling - continued Markov Random Field models
Presentation transcript:

ML estimation in Gaussian graphical models Caroline Uhler Department of Statistics UC Berkeley AMS sectional meeting, Lexington March 27, 2010

Outline Background and setup: - Gaussian graphical models - Maximum likelihood estimation Existence of the maximum likelihood estimate (MLE) “Cone” problem “Probability” problem ML degree of a graph Example:

Gaussian graphical models - undirected graph with - covariance matrix on - concentration matrix with Gaussian graphical model: Data: - i.i.d samples from - sample covariance matrix - sufficient statistics

Geometry Concentration matrices: Covariance matrices:

Example K2,3 5 4 The corresponding Gaussian graphical model consists of multivariate Gaussians with concentration matrix of the form Given a sample covariance matrix the sufficient statistics are 3 1 2

Maximum likelihood estimation Log-likelihood function: Theorem (regular exponential families): In a Gaussian graphical model the MLEs and exist if and only if i.e. is PD-completable. Then is uniquely determined by

Geometry Concentration matrices: Covariance matrices:

Cones Cone of concentration matrices: Cone of sufficient statistics: where respectively

Geometry Concentration matrices: Covariance matrices:

Cones and maximum likelihood estimation Theorem (Sturmfels & U., 2010): is the convex dual to Furthermore, and are closed convex cones which are dual to each other with and Theorem (exponential families): The map is a homeomorphism between and The inverse map takes the sufficient statistics to the MLE of the concentration matrix. Here, is the unique maximizer of the determinant over

Geometry Concentration matrices: Covariance matrices:

Existence of MLE: 2 Problems Given a graph ? Under what conditions on does the MLE exist? (i.e. describe ) “Cone” problem ? Under what conditions on does the MLE exist? “Probability” problem

Problem 1: Example K2,3 Example: 5 4 3 1 2

Problem 1: Example K2,3

Problem 1: K2,m Theorem (U.): The MLE exists for if and only if

Existence of MLE: 2 Problems Given a graph ? Under what conditions on does the MLE exist? ? Under what conditions on does the MLE exist? And with what probability?

Probability of existence Reminder: MLE exists Assume: have length 1 and Note: Existence of the MLE is invariant under a) Rescaling: where is diagonal. b) Orthogonal transformation: where is orthogonal.

Problem 2: K2,m Theorem (U.): The MLE exists on with probability 1 for and does not exist for For let ( ). The MLE exists if and only if lie between and or lie outside and This happens with prob.

ML-degree of a graph The maximum likelihood degree of a statistical model is the number of complex solutions to the likelihood equations for generic data. Generic data: The number of solutions is a constant for all data, except possibly for a lower-dimensional subset of the data space.

A Click to add title ML-degree of a graph Theorem (Sturmfels & U., 2010): chordal if and only if Conjecture (Drton, Sullivant & Sturmfels, 2009): The ML-degree of an m-cycle is given by Click to add title Theorem (U.): The ML-degree of the bipartite graph is given by

Sturmfels & U.: Multivariate Gaussians, semidefinite matrix completion, and convex algebraic geometry (to appear in AISM) U.: Maximum likelihood estimation in Gaussian graphical models (in progress) Thank you!