Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Environmental Data Analysis with MatLab
Lecture 1 Describing Inverse Problems. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability.
Lecture 20 Continuous Problems Linear Operators and Their Adjoints.
Environmental Data Analysis with MatLab Lecture 21: Interpolation.
Lecture 10 Nonuniqueness and Localized Averages. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 14 Nonlinear Problems Grid Search and Monte Carlo Methods.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Environmental Data Analysis with MatLab Lecture 8: Solving Generalized Least Squares Problems.
Lecture 13 L1 , L∞ Norm Problems and Linear Programming
Lecture 23 Exemplary Inverse Problems including Earthquake Location.
Lecture 22 Exemplary Inverse Problems including Filter Design.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Lecture 18 Varimax Factors and Empircal Orthogonal Functions.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Lecture 7 Backus-Gilbert Generalized Inverse and the Trade Off of Resolution and Variance.
Lecture 3 Probability and Measurement Error, Part 2.
Environmental Data Analysis with MatLab Lecture 23: Hypothesis Testing continued; F-Tests.
Environmental Data Analysis with MatLab
Visual Recognition Tutorial
Lecture 2 Probability and Measurement Error, Part 1.
Lecture 5 A Priori Information and Weighted Least Squared.
Lecture 19 Continuous Problems: Backus-Gilbert Theory and Radon’s Problem.
Lecture 15 Nonlinear Problems Newton’s Method. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 16 Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals.
Lecture 17 Factor Analysis. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 21 Continuous Problems Fréchet Derivatives.
Lecture 24 Exemplary Inverse Problems including Vibrational Problems.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 8 Advanced Topics in Least Squares - Part Two -
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
Lecture 3 Review of Linear Algebra Simple least-squares.
Lecture 12 Equality and Inequality Constraints. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 3: Probability and Measurement Error.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Lecture 8 The Principle of Maximum Likelihood. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 11 Vector Spaces and Singular Value Decomposition.
Linear and generalised linear models
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
Lecture II-2: Probability Review
Today Wrap up of probability Vectors, Matrices. Calculus
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
GEO7600 Inverse Theory 09 Sep 2008 Inverse Theory: Goals are to (1) Solve for parameters from observational data; (2) Know something about the range of.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Curve-Fitting Regression
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Lecture 16 - Approximation Methods CVEN 302 July 15, 2002.
Machine Learning CUNY Graduate Center Lecture 2: Math Primer.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Computacion Inteligente Least-Square Methods for System Identification.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2015 Professor Brandon A. Jones Lecture 19: Examples with the Batch Processor.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Filtering and State Estimation: Basic Concepts
Linear regression Fitting a straight line to observations.
12. Principles of Parameter Estimation
Presentation transcript:

Lecture 9 Inexact Theories

Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and Measurement Error, Part 2 Lecture 04The L 2 Norm and Simple Least Squares Lecture 05A Priori Information and Weighted Least Squared Lecture 06Resolution and Generalized Inverses Lecture 07Backus-Gilbert Inverse and the Trade Off of Resolution and Variance Lecture 08The Principle of Maximum Likelihood Lecture 09Inexact Theories Lecture 10Nonuniqueness and Localized Averages Lecture 11Vector Spaces and Singular Value Decomposition Lecture 12Equality and Inequality Constraints Lecture 13L 1, L ∞ Norm Problems and Linear Programming Lecture 14Nonlinear Problems: Grid and Monte Carlo Searches Lecture 15Nonlinear Problems: Newton’s Method Lecture 16Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals Lecture 17Factor Analysis Lecture 18Varimax Factors, Empirical Orthogonal Functions Lecture 19Backus-Gilbert Theory for Continuous Problems; Radon’s Problem Lecture 20Linear Operators and Their Adjoints Lecture 21Fréchet Derivatives Lecture 22 Exemplary Inverse Problems, incl. Filter Design Lecture 23 Exemplary Inverse Problems, incl. Earthquake Location Lecture 24 Exemplary Inverse Problems, incl. Vibrational Problems

Purpose of the Lecture Discuss how an inexact theory can be represented Solve the inexact, linear Gaussian inverse problem Use maximization of relative entropy as a guiding principle for solving inverse problems Introduce F-test as way to determine whether one solution is “better” than another

Part 1 How Inexact Theories can be Represented

How do we generalize the case of an exact theory to one that is inexact?

d obs model, m datum, d m ap m est d pre d=g(m) exact theory case theory

d obs model, m datum, d m ap m est d pre d=g(m) to make theory inexact... must make the theory probabilistic or fuzzy

d obs datum, d model, m m ap d obs m ap d obs m ap m est d pre model, m datum, d combinationtheorya prior p.d.f.

how do you combine two probability density functions ?

so that the information in them is combined...

desirable properties order shouldn’t matter combining something with the null distribution should leave it unchanged combination should be invariant under change of variables

Answer

d obs datum, d model, m m ap d obs m ap d obs m ap m est d pre model, m datum, d theory, p g d obs datum, d model, m m ap d obs m ap d obs m ap m est d pre model, m datum, d (F)(E)(D) a priori, p A total, p T

“solution to inverse problem” maximum likelihood point of (with p N ∝constant) simultaneously gives m est and d pre

probability that the estimated model parameters are near m and the predicted data are near d probability that the estimated model parameters are near m irrespective of the value of the predicted data T

conceptual problem do not necessarily have maximum likelihood points at the same value of m and T

d obs d pre model, m datum, d m ap m est model, m m est ’ p(m)

illustrates the problem in defining a definitive solution to an inverse problem

illustrates the problem in defining a definitive solution to an inverse problem fortunately if all distributions are Gaussian the two points are the same

Part 2 Solution of the inexact linear Gaussian inverse problem

Gaussian a priori information

a priori values of model parameters their uncertainty

Gaussian observations

observed data measurement error

Gaussian theory

linear theory uncertainty in theory

mathematical statement of problem find (m,d) that maximizes p T (m,d) = p A (m) p A (d) p g (m,d) and, along the way, work out the form of p T (m,d)

notational simplification group m and d into single vector x = [d T, m T ] T group [cov m] A and [cov d] A into single matrix write d-Gm=0 as Fx=0 with F=[I, –G]

after much algebra, we find p T (x) is a Gaussian distribution with mean and variance

after much algebra, we find p T (x) is a Gaussian distribution with mean and variance solution to inverse problem

after pulling m est out of x*

reminiscent of G T (GG T ) -1 minimum length solution

after pulling m est out of x* error in theory adds to error in data

after pulling m est out of x* solution depends on the values of the prior information only to the extent that the model resolution matrix is different from an identity matrix

and after algebraic manipulation which also equals reminiscent of (G T G) -1 G T least squares solution

interesting aside weighted least squares solution is equal to the weighted minimum length solution

what did we learn? for linear Gaussian inverse problem inexactness of theory just adds to inexactness of data

Part 3 Use maximization of relative entropy as a guiding principle for solving inverse problems

from last lecture

assessing the information content in p A (m) Do we know a little about m or a lot about m ?

Information Gain, S - S called Relative Entropy

m p A (m) S(σ A ) p N (m) σAσA (A) (B)

Principle of Maximum Relative Entropy or if you prefer Principle of Minimum Information Gain

find solution p.d.f. p T (m) that has smallest possible new information as compared to a priori p.d.f. p A (m) find solution p.d.f. p T (m) that has the largest relative entropy as compared to a priori p.d.f. p A (m) or if you prefer

properly normalized p.d.f. data is satisfied in the mean or expected value of error is zero

After minimization using Lagrange Multipliers process p T (m) is Gaussian with maximum likelihood point m est satisfying

After minimization using Lagrane Multipliers process p T (m) is Gaussian with maximum likelihood point m est satisfying just the weighted minimum length solution

What did we learn? Only that the Principle of Maximum Entropy is yet another way of deriving the inverse problem solutions we are already familiar with

Part 4 F-test as way to determine whether one solution is “better” than another

Common Scenario two different theories solution m est A M A model parameters prediction error E A solution m est B M B model parameters prediction error E B

Suppose E B < E A Is B really better than A ?

What if B has many more model parameters than A M B >> M A Is B fitting better any surprise?

Need to against Null Hypothesis The difference in error is due to random variation

suppose error e has a Gaussian p.d.f. uncorrelated uniform variance σ d

estimate variance

want to known the probability density function of

actually, we’ll use the quantity which is the same, as long as the two theories that we’re testing is applied to the same data

p(F N,2 ) p(F N,5 ) p(F N,50 ) F F F F p(F N,25 ) N=2 50 N=250 N=2 50 N=2 50 p.d.f. of F is known

as is its mean and variance

example same dataset fit with a straight line and a cubic polynomial

(A) Linear fit, N-M=9, E=0.030 (B) Cubic fit, N-M=7, E=0.006 zizi zizi didi didi

(A) Linear fit, N-M=9, E=0.030 (B) Cubic fit, N-M=7, E=0.006 zizi zizi didi didi F 7,9 = 4.1 est

probability that F >F est (cubic fit seems better than linear fit) by random chance alone or F < 1/F est (linear fit seems better than cubic fit) by random chance alone

in MatLab P = 1 - (fcdf(Fobs,vA,vB)-fcdf(1/Fobs,vA,vB));

answer: 6% The Null Hypothesis that the difference is due to random variation cannot be rejected to 95% confidence