G(m)=d mathematical model d data m model G operator d=G(m true )+  = d true +  Forward problem: find d given m Inverse problem (discrete parameter estimation):

Slides:



Advertisements
Similar presentations
Environmental Data Analysis with MatLab Lecture 8: Solving Generalized Least Squares Problems.
Advertisements

Lecture 13 L1 , L∞ Norm Problems and Linear Programming
Lecture 23 Exemplary Inverse Problems including Earthquake Location.
Inversion Transforming the apparent to « real » resistivity. Find a numerical model that explains the field measurment.
1 Functions and Applications
Lecture 22 Exemplary Inverse Problems including Filter Design.
Lecture 7 Backus-Gilbert Generalized Inverse and the Trade Off of Resolution and Variance.
COMP 116: Introduction to Scientific Programming Lecture 11: Linear Regression.
Lecture 3 Review of Linear Algebra Simple least-squares.
Inverse Theory CIDER seismology lecture IV July 14, 2014
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Lecture 5 A Priori Information and Weighted Least Squared.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 21 Continuous Problems Fréchet Derivatives.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 8 Advanced Topics in Least Squares - Part Two -
Newton’s Method applied to a scalar function Newton’s method for minimizing f(x): Twice differentiable function f(x), initial solution x 0. Generate a.
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
Lecture 3 Review of Linear Algebra Simple least-squares.
Curve-Fitting Regression
Lecture 12 Equality and Inequality Constraints. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 7 Advanced Topics in Least Squares. the multivariate normal distribution for data, d p(d) = (2  ) -N/2 |C d | -1/2 exp{ -1/2 (d-d) T C d -1 (d-d)
Lecture 4: Practical Examples. Remember this? m est = m A + M [ d obs – Gm A ] where M = [G T C d -1 G + C m -1 ] -1 G T C d -1.
Lecture 8 The Principle of Maximum Likelihood. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
1 MF-852 Financial Econometrics Lecture 2 Matrix Operations in Econometrics, Optimization with Excel Roy J. Epstein Fall 2003.
Newton’s Method applied to a scalar function Newton’s method for minimizing f(x): Twice differentiable function f(x), initial solution x 0. Generate a.
1 Neural Nets Applications Vectors and Matrices. 2/27 Outline 1. Definition of Vectors 2. Operations on Vectors 3. Linear Dependence of Vectors 4. Definition.
Lecture 11 Vector Spaces and Singular Value Decomposition.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Linear and generalised linear models
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Classification and Prediction: Regression Analysis
Chapter 10 Real Inner Products and Least-Square (cont.)
Geophysical Inverse Problems with a focus on seismic tomography CIDER2012- KITP- Santa Barbara.
Simultaneous inversion of seabed and water column sound speed profiles in range-dependent shallow-water environments Megan S. Ballard and Kyle M. Becker.
Making Models from Data A Basic Overview of Parameter Estimation and Inverse Theory or Four Centuries of Linear Algebra in 10 Equations Rick Aster Professor.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Scientific Computing Linear Least Squares. Interpolation vs Approximation Recall: Given a set of (x,y) data points, Interpolation is the process of finding.
ENE MATHLAB ® Lecture 3: Matrix Review. Determinant  To use MATLAB to compute determinants:  Enter the determinant as an array.  Use ‘det’ command.
Curve-Fitting Regression
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 26: Singular Value Decomposition.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 4. Least squares.
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
Lecture 16 - Approximation Methods CVEN 302 July 15, 2002.
Ridge Regression: Biased Estimation for Nonorthogonal Problems by A.E. Hoerl and R.W. Kennard Regression Shrinkage and Selection via the Lasso by Robert.
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
Inverse Modeling of Surface Carbon Fluxes Please read Peters et al (2007) and Explore the CarbonTracker website.
CS654: Digital Image Analysis Lecture 22: Image Restoration.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Problems in solving generic AX = B Case 1: There are errors in data such that data cannot be fit perfectly (analog: simple case of fitting a line while.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Geology 5670/6670 Inverse Theory 4 Feb 2015 © A.R. Lowry 2015 Read for Fri 6 Feb: Menke Ch 4 (69-88) Last time: The Generalized Inverse The Generalized.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2015 Professor Brandon A. Jones Lecture 26: Cholesky and Singular Value.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
1 SYSTEM OF LINEAR EQUATIONS BASE OF VECTOR SPACE.
Geology 5670/6670 Inverse Theory 6 Feb 2015 © A.R. Lowry 2015 Read for Mon 9 Feb: Menke Ch 5 (89-114) Last time: The Generalized Inverse; Damped LS The.
Lecture 2 Linear Inverse Problems and Introduction to Least Squares.
ASEN 5070: Statistical Orbit Determination I Fall 2014
Ch3: Model Building through Regression
Some useful linear algebra
Linear regression Fitting a straight line to observations.
Discrete Least Squares Approximation
Mathematical Foundations of BME Reza Shadmehr
Presentation transcript:

G(m)=d mathematical model d data m model G operator d=G(m true )+  = d true +  Forward problem: find d given m Inverse problem (discrete parameter estimation): find m given d Discrete linear inverse problem: Gm=d

G(m)=d mathematical model Discrete linear inverse problem: Gm=d Method of Least Squares: Minimize E=∑ e i 2 = ∑ (d i obs -d i pre ) 2 o o o o o e i d i pre D i obs z i

E=e T e=(d-Gm) T (d-Gm) =∑ [d i -∑G ij m j ] [d i -∑G ik m k ] =∑ ∑ m j m k ∑ G ij G ik -2∑ m j ∑ G ij d i + ∑ d i d i ∂/∂m q [∑ ∑ m j m k ∑ G ij G ik ] = ∑ ∑ [  jq m k +m j  kq ]∑G ij G ik = 2 ∑ m k ∑ G iq G ik -2 ∂/∂m q [∑ m j ∑ G ij d i ] = -2∑  jq ∑G ij d i = -2∑ G iq d i ∂/∂m q [∑d i d i ]=0 i j k j k i j i i j k i k i j I j i i i

∂/∂m q = 0 = 2 ∑ m k ∑ G iq G ik - 2∑ G iq d i In matrix notation: G T Gm - G T d = 0 m est = [G T G] -1 G T d assuming [G T G] -1 exists This is the least squares solution to Gm=d k i i

Example of fitting a straight line m est = [G T G] -1 G T d assuming [G T G] -1 exists m ∑ x i ∑ x i ∑ x i … 1 1 x 1 [G T G] = 1 x 2 = x 1 x 2.. x m. 1 x m m ∑ x i -1 ∑ x i ∑ x i 2 [G T G] -1 =

Example of fitting a straight line m est = [G T G] -1 G T d assuming [G T G] -1 exists ∑ d i ∑ x i d i 1 1 … 1 d 1 [G T d] = d 2 = x 1 x 2.. x m. d m m ∑ x i -1 ∑ x i ∑ x i 2 [G T G] -1 G T d = ∑ d i ∑ x i d i

The existence of the Least Squares Solution m est = [G T G] -1 G T d assuming [G T G] -1 exists Consider the straight line problem with only 1 data point o ? m ∑ x i -1 1 x 1 -1 [G T G] -1 = = ∑ x i ∑ x i 2 x 1 x 1 2 The inverse of a matrix is proportional to the reciprocal of the determinant of the matrix, i.e., [G T G] -1  1/(x 1 2 -x 1 2 ), which is clearly singular, and the formula for the least squares fails.

Classification of inverse problems: Over-determined Under-determined Mixed-determined Even-determined

Over-determined problems: Too much information contained in Gm=d to possess an exact solution … Least squares gives a ‘best’ approximate solution.

Even-determined problems: Exactly enough information to determine the model parameters. There is only one solution and it has zero prediction error

Under-determined Problems: Mixed-determined problems - non-zero prediction error Purely underdetermined problems - zero prediction error

Purely Under-determined Problems: # of parameters > # of equations Possible to find more than 1 solution with 0 prediction error (actually infinitely many) To obtain a solution, we must add some information not contained in Gm=d : a priori information Example: Fitting a straight line through a single data point, we may require that the line passes through the origin Common a priori assumption: Simple model solution best. Measure of simplicity could be Euclidian length, L=m T m = ∑ m i 2

Purely Under-determined Problems: Problem: Find the m est that minimizes L=m T m = ∑ m i 2 subject to the constraint that e=d-Gm=0  (m)= L+∑ i e i = ∑ m i 2 +∑ i [ d i - ∑ G ij m j ] ∂  (m)/∂m q = 2 ∑ m i ∂m i /∂m q -∑ i ∑ G ij ∂m j /∂m q ] = 2m q - ∑ i G iq = 0 In matrix notation: 2m = G T (1), along with Gm=d (2) Inserting (1) into (2) we get d=Gm=G[G T /2],  = 2[GG T ] -1 d and inserting into (1): m = G T [GG T ] -1 d - solution exist when purely underdetermined

Mixed-determined problems Over Under Mixed determined determined determined 1)Partition into overdetermined and underdetermined parts, solve by LS and minimum norm - SVD (later) 2)Minimize some combination of the prediction error and solution length for the unpartitioned model  (m)=E+  2 L=e T e+  2 m T m m est =[G T G+  2 I] -1 G T d - damped least squares

Mixed-determined problems  (m)=E+  2 L=e T e+  2 m T m m est =[G T G+  2 I] -1 G T d - damped least squares  Regularization parameter 0th-order Tikhonov Regularization ||m||  ||Gm-d|| Min ||m|| 2, ||Gm-d|| 2 <  min ||Gm-d|| 2, ||m|| 2 <  ||m||  ‘L-curves’ ||Gm-d||

Other A Priori Info: Weighted Least Squares Data weighting (weighted measures if prediction error) E=e T W e e W e is a weighting matrix, defining relative contribution of each individual error to the total prediction error (usually diagonal). For example, for 5 observations, the 3rd may be twice as accurately determined as the others: Diag(W e )=[1, 1, 2, 1, 1] T Completely overdetermined problem: m est =[G T W e G] -1 G T W e d

Other A Priori Info: Constrained Regression d i =m 1 +m 2 x i Constraint: line must pass through (x’,d’): d’=m 1 +m 2 x’ Fm= [1 x’] [m 1 m 2 ] T = [d’] Similar to the unconstrained solution (2.5) we get: m 1 est M ∑ x i 1 -1 ∑ d i m 2 est = ∑ x i ∑ x i 2 x’ ∑ x i d i  1 1 x’ 0 d’ o o o o (x’,d’) o d x M ∑ x i -1 ∑ x i ∑ x i 2 Unconstrained solution: [G T G] -1 G T d = ∑ d i ∑ x i d i

Other A Priori Info: Weighting model parameters Instead of using minimum length as solution simplicity, One may impose smoothness in the model: -1 1 m m 2 l =... = Dm m N D is the flatness matrix L=l T l=[Dm] T [Dm]=m T D T Dm=m T W m m, W m =D T D firsth-order Tikhonov Regularization - min||Gm-d||   ||Lm|| 2 2

Other A Priori Info: Weighting model parameters Instead of using minimum length as solution simplicity, One may impose smoothness in the model: m m 2 l =.... = Dm m N D is the roughness matrix L=l T l=[Dm] T [Dm]=m T D T Dm=m T W m m, W m =D T D 2nd-order Tikhonov Regularization- min||Gm-d||   ||Lm|| 2 2