The Multivariate Normal Distribution, Part I

Slides:



Advertisements
Similar presentations
Component Analysis (Review)
Advertisements

Multivariate distributions. The Normal distribution.
II. The Multivariate Normal Distribution “…it is not enough to know that a sample could have come from a normal population; we must be clear that it is.
Sampling Distributions
Multivariate Distance and Similarity Robert F. Murphy Cytometry Development Workshop 2000.
Probability theory 2011 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different definitions.
CS Pattern Recognition Review of Prerequisites in Math and Statistics Prepared by Li Yang Based on Appendix chapters of Pattern Recognition, 4.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Continuous Random Variables and Probability Distributions
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Statistical Intervals Based on a Single Sample.
Techniques for studying correlation and covariance structure
Random Variable and Probability Distribution
Lecture II-2: Probability Review
The Multivariate Normal Distribution, Part 1 BMTRY 726 1/10/2014.
The Multivariate Normal Distribution, Part 2 BMTRY 726 1/14/2014.
1 Multivariate Normal Distribution Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Modern Navigation Thomas Herring
Separate multivariate observations
The Neymann-Pearson Lemma Suppose that the data x 1, …, x n has joint density function f(x 1, …, x n ;  ) where  is either  1 or  2. Let g(x 1, …,
Principles of Pattern Recognition
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
Ch 2. Probability Distributions (1/2) Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by Yung-Kyun Noh and Joo-kyung Kim Biointelligence.
Chapter 5.6 From DeGroot & Schervish. Uniform Distribution.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
1 Sample Geometry and Random Sampling Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
The Multivariate Gaussian
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 8: Estimating with Confidence Section 8.1 Confidence Intervals: The.
Sampling and Sampling Distributions
SUR-2250 Error Theory.
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Basic simulation methodology
LECTURE 10: DISCRIMINANT ANALYSIS
STATISTICS Random Variables and Distribution Functions
Inference for the mean vector
3.1 Expectation Expectation Example
CONCEPTS OF ESTIMATION
ECE 417 Lecture 4: Multivariate Gaussians
Presented by Kojo Essuman Ackah Spring 2018 STA 6557 Project
The Multivariate Normal Distribution, Part 2
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Chapter 8: Estimating with Confidence
Principal Components What matters most?.
LECTURE 09: DISCRIMINANT ANALYSIS
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter-1 Multivariate Normal Distributions
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8: Estimating with Confidence
Chapter 8 Estimation.
Chapter 8: Estimating with Confidence
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Regression Models - Introduction
Presentation transcript:

The Multivariate Normal Distribution, Part I BMTRY 726 5/17/2018

Univariate Normal Recall the density function of the univariate normal We can rewrite this as

Multivariate Normal Distribution We denote the MVN distribution as What is the density function of X?

Multivariate Normal Distribution What is the density function of X?

Multivariate Normal Note, the density does not exist if S is not positive definite = 0 does not exist We will assume that is positive definite for most of the MVN methods we discuss

Multivariate Density Function If we assume that S is positive definite is the square of the generalized distance from x to m. Also called Squared statistical distance of x to m. Squared Mahalanobis distance of x to m Squared standardized distance of x to m

Why Multivariate Normal The MVN distribution makes a good choice in statistics for several reasons Mathematical simplicity Multivariate central limit theorem Many naturally occurring phenomenon approximately exhibit this distribution

Bivariate Normal Example Consider samples from Let’s write out the joint distribution of x1 and x2

Bivariate Normal Example Joint distribution of x1 and x2

Bivariate Normal Example Joint distribution of x1 and x2

Bivariate Normal Example This yields joint distribution of x1 and x2 in the form

Bivariate Normal Example The density if a function of m1, m2, s1, s2, and r The density is well defined if -1 < r < 1 If r = 0, then …

Examples of Bivariate Normal

R Code to Play Around with MVN If you want to vary parameters and see what changes… > library(MASS) > S<-matrix(c(1, 0.5, 0.5, 1), 2) > bivn <- mvrnorm(100000, mu = c(0, 0), Sigma = S) > > # now we do a kernel density estimate > bivn.kde <- kde2d(bivn[,1], bivn[,2], n = 50) > persp(bivn.kde, phi = 45, theta = 30, xlab="X1", ylab="X2", zlab="phi") > # alternatively, use a “fancy” perspective > persp(bivn.kde, phi = 45, theta = 30, shade = .2, border = NA)

Contours of constant density What if we take a slice of this bivariate distribution at a constant height? i.e.

Contours of constant density The density is constant for all points for which This is an equation for an ellipse centered at

Contours of constant density What happens when

Bivariate Normal Example Let’s look at an example of the bivariate normal when we vary some of the parameters…

Examples X2 X2 X1 X1 X2 X2 X1 X1

Contours of constant density What about when

Contours of constant density How do we find the axes of the ellipse? Axes are in the direction of the eigenvectors of S-1 Axes lengths are proportional to the reciprocals of the square root of the eigenvalues of S-1 We can get these from S (avoid calculating S-1) Let’s look at this for the bivariate case... We must find the eigenvalues and eigenvectors for S Eigenvalues: Eigenvectors:

Eigenvalues of S:

Eigenvalues of S:

The corresponding eigenvector, e1, of S:

The corresponding eigenvector, e1, of S:

Similarly we can find e2, which corresponds to l2 : The axes of the contours of constant density will have length -

If we let then are the eigenvalues of S and e1 and e2 are the corresponding eigenvectors

The ratio of the lengths of the axes The actual lengths depend on the contour being considered. For the (1-a)x100% contour, the ½ lengths are given by Thus the solid ellipsoid of x values satisfying has probability 1-a. -

Univariate case: length of the interval containing the central 95% of the population is proportional to s Bivariate case: the area of the region containing 95% of the population is proportional to .

The “area” of this smallest ellipse in the 2-D case is: We can call this “smallest” region the central (1-a)x100% of the multivariate normal population. The “area” of this smallest ellipse in the 2-D case is: This extends to higher dimensions (think volume) Consider The smallest region for which there is 1-a that a randomly selected observation falls in the region is a p-dimensional ellipsoid centered at m with volume

Visual of the 3-dimensional case

Why Multivariate Normal Recall, statisticians like the MVN distribution because… Mathematically simple Multivariate central limit theorem applies Natural phenomena are often well approximated by a MVN distribution So what are some “fun” mathematical properties that make is so nice?

Properties of MVN Result 4.2: If then has a univariate normal distribution with mean and variance

Example

Properties of MVN Result 4. 3: Any linear transformation of a multivatiate normal random vector has a normal distribution So if and and B is a k x p matrix of constants then

Result 4.3 In the development of univariate statistics, we often transform our observed r.v. to the canonical form: Result 4.3 is of particular interest because we can use it to derive the canonical form of the multivariate normal… But first, we need to consider a specific matrix decomposition

Spectral Decomposition Given S is a non-negative definite, symmetric, real matrix, then S can be decomposed according to: Where the eigenvalues are The eigenvectors of S are e1, e2,...,ep And these satisfy the expression

Marginal Distributions Result 4.4: Consider subsets of Xi’s in X. These subsets are also distributed (multivariate) normal. If Then the marginal distributions of X1 and X2 is:

Example Consider , find the marginal distribution of the 1st and 3rd components

Example Consider , find the marginal distribution of the 1st and 3rd components

Marginal Distributions cont’d The converse of result 4.4 is not always true, an additional assumption is needed. Result 4.5(c): If… and X1 is independent of X2 then

Result 4. 5(a): If X1(qx1) and X2(p-qx1) are independent then Cov(X1,X2) = 0 (b) If Then X1(qx1) and X2(p-qx1) are independent iff

Example

Next Time More properties of the MVN and sampling from MVN…