Environmental Data Analysis with MatLab Lecture 7: Prior Information.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

The Maximum Likelihood Method
Environmental Data Analysis with MatLab Lecture 10: Complex Fourier Series.
Environmental Data Analysis with MatLab
Environmental Data Analysis with MatLab Lecture 21: Interpolation.
Lecture 10 Nonuniqueness and Localized Averages. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 15: Factor Analysis.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
General Linear Model With correlated error terms  =  2 V ≠  2 I.
Environmental Data Analysis with MatLab Lecture 8: Solving Generalized Least Squares Problems.
Lecture 13 L1 , L∞ Norm Problems and Linear Programming
Environmental Data Analysis with MatLab Lecture 9: Fourier Series.
Environmental Data Analysis with MatLab
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Environmental Data Analysis with MatLab Lecture 13: Filter Theory.
Environmental Data Analysis with MatLab Lecture 16: Orthogonal Functions.
Lecture 3 Probability and Measurement Error, Part 2.
Environmental Data Analysis with MatLab Lecture 23: Hypothesis Testing continued; F-Tests.
Environmental Data Analysis with MatLab Lecture 11: Lessons Learned from the Fourier Transform.
Environmental Data Analysis with MatLab
Environmental Data Analysis with MatLab Lecture 12: Power Spectral Density.
Lecture 2 Probability and Measurement Error, Part 1.
Lecture 5 A Priori Information and Weighted Least Squared.
Environmental Data Analysis with MatLab Lecture 17: Covariance and Autocorrelation.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 8 Advanced Topics in Least Squares - Part Two -
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
Lecture 3 Review of Linear Algebra Simple least-squares.
Lecture 7 Advanced Topics in Least Squares. the multivariate normal distribution for data, d p(d) = (2  ) -N/2 |C d | -1/2 exp{ -1/2 (d-d) T C d -1 (d-d)
Lecture 5 Probability and Statistics. Please Read Doug Martinson’s Chapter 3: ‘Statistics’ Available on Courseworks.
Environmental Data Analysis with MatLab Lecture 3: Probability and Measurement Error.
Lecture 8 The Principle of Maximum Likelihood. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 24: Confidence Limits of Spectra; Bootstraps.
Lecture 11 Vector Spaces and Singular Value Decomposition.
Linear and generalised linear models
Linear and generalised linear models
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
Lecture II-2: Probability Review
Introduction to Regression Analysis, Chapter 13,
Environmental Data Analysis with MatLab Lecture 20: Coherence; Tapering and Spectral Analysis.
Principles of Pattern Recognition
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
Environmental Data Analysis with MatLab Lecture 10: Complex Fourier Series.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Computacion Inteligente Least-Square Methods for System Identification.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 14: Applications of Filters.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Probability Theory and Parameter Estimation I
CH 5: Multivariate Methods
Special Topics In Scientific Computing
Lecture 26: Environmental Data Analysis with MatLab 2nd Edition
Environmental Data Analysis with MatLab 2nd Edition
Environmental Data Analysis with MatLab 2nd Edition
Computing and Statistical Data Analysis / Stat 8
Regression Lecture-5 Additional chapters of mathematics
OVERVIEW OF LINEAR MODELS
Environmental Data Analysis with MatLab
Presentation transcript:

Environmental Data Analysis with MatLab Lecture 7: Prior Information

Lecture 01Using MatLab Lecture 02Looking At Data Lecture 03Probability and Measurement Error Lecture 04Multivariate Distributions Lecture 05Linear Models Lecture 06The Principle of Least Squares Lecture 07Prior Information Lecture 08Solving Generalized Least Squares Problems Lecture 09Fourier Series Lecture 10Complex Fourier Series Lecture 11Lessons Learned from the Fourier Transform Lecture 12Power Spectra Lecture 13Filter Theory Lecture 14Applications of Filters Lecture 15Factor Analysis Lecture 16Orthogonal functions Lecture 17Covariance and Autocorrelation Lecture 18Cross-correlation Lecture 19Smoothing, Correlation and Spectra Lecture 20Coherence; Tapering and Spectral Analysis Lecture 21Interpolation Lecture 22 Hypothesis testing Lecture 23 Hypothesis Testing continued; F-Tests Lecture 24 Confidence Limits of Spectra, Bootstraps SYLLABUS

purpose of the lecture understand the advantages and limitations of supplementing observations with prior information

when least-squares fails

x d x1x1 x d x*x* d1d1 d*d* fitting of straight line cases were there’s more than one solution E exactly 0 for any lines passing through point one point E minimum for all lines passing through point x*

when determinant of [G T G] is zero that is, D=0 [G T G] -1 is singular

x d x1x1 d1d1 N=1 case E exactly 0 for any lines passing through point one point

x d x*x* d*d* x i =x* case E minimum for any lines passing through point x*

least-squares fails when the data do not uniquely determine the solution

if [G T G] -1 is singular least squares solution doesn’t exist if [G T G] -1 is almost singular least squares is useless because it has high variance very large

guiding principle for avoiding failure add information to the problem that guarantees that matrices like [G T G] are never singular such information is called prior information

examples of prior information soil has density will be around 1500 kg/m 3 give or take 500 or so chemical components sum to 100% pollutant transport is subject to the diffusion equation water in rivers always flows downhill

prior information things we know about the solution based on our knowledge and experience but not directly based on data

simplest prior information m is near some value, m m ≈ m with covariance C m p

use Normal p.d.f. to prepresent prior information

m1m1 m2m pp(m)pp(m) prior information example: m 1 = 10 ± 5 m 2 = 20 ± 5 m 1 and m 2 uncorrelated

Normal p.d.f. defines an “error in prior information” individual errors weighted by their certainty

linear prior information with covariance C h

example relevant to chemical constituents Hh

use Normal p.d.f. to represent prior information

Normal p.d.f. defines an “error in prior information” individual errors weighted by their certainty

(Technically, the p.d.f.’s are only proportional when the Jacobian Determinant is constant, which it is in this case). p p (m) ∝ so we can view this formula as a p.d.f. for the model parameters, m since m ∝ h p p (m) ∝ p p (h)

now suppose that we observe some data: d = d obs with covariance C d

use Normal p.d.f. to represent the observations d with covariance C d

now assume that the mean of the data is predicted by the model: d = Gm

represent the observations with a Normal p.d.f. mean of data predicted by the model observations weighted by their certainty p(d) =

Normal p.d.f. defines an “error in data” p(d) = weighted least- squares error

think of p(d) as a conditional p.d.f. probability that a particular set of data values will be observed given a particular choice of model parameters

m2m p(d|m) m1m1 example: one datum 2 model parameters model d 1 =m 1 –m 2 one observation d 1 obs = 0 ± 3

now use Bayes theorem to update the prior information with the observations

ignore for a moment

Bayes Theorem in words

so the updated p.d.f. for the model parameters is: data partprior information part

this p.d.f. defines a “total error” weighted least squares error in the data weighted error in the prior information with

Generalized Principle of Least Squares the best m est is the one that minimizes the total error with respect to m which is the same one as the one that maximized p(m|d) with respect to m

m1m1 m2m m2m m2m A) p p (m) B) p(d|m)C) p(m|d) m1m1 m1m1 continuing the example … best estimate of the model parameters

generalized least squares find the m that minimizes

generalized least squares solution pattern same as ordinary least squares but with more complicated matrices