The Basics of Inversion

Slides:



Advertisements
Similar presentations
Environmental Data Analysis with MatLab
Advertisements

Lecture 1 Describing Inverse Problems. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability.
Introduction: Correlation and Regression The General Linear Model is a phrase used to indicate a class of statistical models which include simple linear.
Environmental Data Analysis with MatLab Lecture 8: Solving Generalized Least Squares Problems.
1 Functions and Applications
P M V Subbarao Professor Mechanical Engineering Department
Lecture (14,15) More than one Variable, Curve Fitting, and Method of Least Squares.
Simple Linear Regression
Lecture 3 Review of Linear Algebra Simple least-squares.
A Short Introduction to Curve Fitting and Regression by Brad Morantz
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Part 1 Cross Sectional Data
2.2 Correlation Correlation measures the direction and strength of the linear relationship between two quantitative variables.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 8 Advanced Topics in Least Squares - Part Two -
Chapter 10 Simple Regression.
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
Lecture 12 Equality and Inequality Constraints. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 7 Advanced Topics in Least Squares. the multivariate normal distribution for data, d p(d) = (2  ) -N/2 |C d | -1/2 exp{ -1/2 (d-d) T C d -1 (d-d)
Statistics for Business and Economics
Lecture 4: Practical Examples. Remember this? m est = m A + M [ d obs – Gm A ] where M = [G T C d -1 G + C m -1 ] -1 G T C d -1.
Lecture 5 Probability and Statistics. Please Read Doug Martinson’s Chapter 3: ‘Statistics’ Available on Courseworks.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
The Simple Regression Model
The Basics of Earthquake Location William Menke Lamont-Doherty Earth Observatory Columbia University.
Lecture 9 Interpolation and Splines. Lingo Interpolation – filling in gaps in data Find a function f(x) that 1) goes through all your data points 2) does.
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
1 Relationships We have examined how to measure relationships between two categorical variables (chi-square) one categorical variable and one measurement.
Correlation & Regression Math 137 Fresno State Burger.
Linear Regression Analysis
Calibration & Curve Fitting
Lecture 3: Bivariate Data & Linear Regression 1.Introduction 2.Bivariate Data 3.Linear Analysis of Data a)Freehand Linear Fit b)Least Squares Fit c)Interpolation/Extrapolation.
Relationship of two variables
MAT 254 – Probability and Statistics Sections 1,2 & Spring.
Making Models from Data A Basic Overview of Parameter Estimation and Inverse Theory or Four Centuries of Linear Algebra in 10 Equations Rick Aster Professor.
Chapter 13 Statistics © 2008 Pearson Addison-Wesley. All rights reserved.
Statistics for Business and Economics Chapter 10 Simple Linear Regression.
© 2008 Pearson Addison-Wesley. All rights reserved Chapter 1 Section 13-6 Regression and Correlation.
Quality of Curve Fitting P M V Subbarao Professor Mechanical Engineering Department Suitability of A Model to a Data Set…..
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
G(m)=d mathematical model d data m model G operator d=G(m true )+  = d true +  Forward problem: find d given m Inverse problem (discrete parameter estimation):
BIOL 582 Lecture Set 11 Bivariate Data Correlation Regression.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
Scatterplot and trendline. Scatterplot Scatterplot explores the relationship between two quantitative variables. Example:
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Lecture 10: Correlation and Regression Model.
Correlation & Regression Analysis
1 Data Analysis Linear Regression Data Analysis Linear Regression Ernesto A. Diaz Department of Mathematics Redwood High School.
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Statistics: Unlocking the Power of Data Lock 5 STAT 250 Dr. Kari Lock Morgan Simple Linear Regression SECTION 2.6 Least squares line Interpreting coefficients.
Seismological Analysis Methods Receiver FunctionsBody Wave Tomography Surface (Rayleigh) wave tomography Good for: Imaging discontinuities (Moho, sed/rock.
Psychology 202a Advanced Psychological Statistics October 22, 2015.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Geology 5670/6670 Inverse Theory 12 Jan 2015 © A.R. Lowry 2015 Read for Wed 14 Jan: Menke Ch 2 (15-37) Last time: Course Introduction (Cont’d) Goal is.
Gravity Data Reduction
Correlation & Regression
Chapter 5 STATISTICS (PART 4).
Discrete Least Squares Approximation
Presentation transcript:

The Basics of Inversion Short Course The Basics of Inversion Bill Menke Columbia University

The power of simple linear models Lectures The power of simple linear models Probability and what it has to do with data analysis 3. Inferences using Least-Squares 4. Examples

MatLab Scripts for all calculations in these lectures www.ldeo.columbia.edu/users/menkecourses.html

The power of simple, linear models Lecture 1 The power of simple, linear models

data, d. … what you measure model parameter, m data, d … what you measure model parameter, m … what you want to know conceptual model … links the two

At the start of a project, spend a few moments identifying. data At the start of a project, spend a few moments identifying data model parameters conceptual model think about the strengths and weaknesses of each

Example – seismic tomography of the earth’s mantle Example – seismic tomography of the earth’s mantle data: traveltimes of shear waves but the seismometer measured wiggles … model parameter: shear velocity in mantle but I really wanted to know temperature shear velocity is just a proxy for temperature conceptual model ray theory, an approximation to how vibrations travel through the earth

example: river flow and rain

watershed with rain and transport data: discharge on a succession of days, d1, d2, d3, … model parameters: rain on a succession of days m1, m2, m3, … conceptual model: watershed with rain and transport di = fcn( mi, mi-1, mi-2, …) Causal: discharge depends only on present and past rain

example: CO2 and combustion of fossil fuels

data: CO2 on a succession of days, data: CO2 on a succession of days, d1, d2, d3, … model parameters: combustion rate on a succession of days m1, m2, m3, … conceptual model: global carbon cycle (transport, storage in biosphere and oceans, etc) di = fcn( mi, mi-1, mi-2, …)

example: gravity anomaly and the earth’s density

data: strength of gravity on a 2D grid of points on the earth’s surface d1, d2, d3, … model parameters: density of the earth on a 3D grid of points in the earth’s interior m1, m2, m3, … conceptual model: Newton’s third law: mass causes gravitational attraction di = fcn(mi+1, mi, mi-1, …) not causal

example: straight line relationship d x

data: d1, d2, d3, … model parameters: the slope and intercept of the line m1, m2 conceptual model: data are on a line: di = m1 + m2 xi or if you prefer di = a + b xi

In all these examples the data are linearly related to the model parameters (at least to first approximation)

Data: d= model parameters: m= Linear model: d = Gm d1 d2 … dN m1 m2 … “data kernel” matrix

Straight line relationship d = G m 1 x1 x2 … xN d1 d2 … dN m1 m2 = Note that the data kernel matrix embodies the “geometry” of the measurements, that is, the x’s at which the d’s were measured

d: gravity anomaly in vertical direction Gravity anomalies d: gravity anomaly in vertical direction m: density anomaly of a small cube of volume Dv di xi mi qij yi gravitational constant Newton’s Inverse-square Law: di = Scubes g Dv mj cosqij/ |xi - yj|2

Gravity anomalies d = G m Newton’s law: zij = g Dv cosqij / |xi - yj|2 … Z1M Z21 Z22 Z2M ZN1 ZN2 ZNM m1 m2 … mM d1 d2 … dN = vertical component of gravity anomaly measured at position xi density anomaly of small cube of volume Dv located at position yi once again, G embodies “geometry” Newton’s law: zij = g Dv cosqij / |xi - yj|2

Thinking About Error error = observed data – predicted data e = dobs – dpre = dobs – Gmest always plot your data and look at the error!

Guess values for a, b ypre = aguess + bguessx Prediction error = observed minus predicted e = dobs - dpre Total error: sum of squared predictions errors E = Σ ei2 = eT e aguess=2.0 bguess=2.4

Systematically examine combinations of (a, b) on a 101101 grid Minimum total error E is here Note E is not zero apre Error Surface bpre

Error Surface Note Emin is not zero Here are best-fitting a, b best-fitting line Error Surface

Note some range of values where the error is about the same as the minimun value, Emin Emin is here Error pretty close to Emin everywhere in here All a’s in this range and b’s in this range have pretty much the same error Error Surface

conclusion the shape of the error surface controls the accuracy by which (a,b) can be estimated

What controls the shape of the error surface What controls the shape of the error surface? Let’s examine effect of increasing the error in the data

The minimum error increases, but the shape of the error surface is pretty much the same Error in data = 0.5 Emin = 0.20 -5 5 Error in data = 5.0 Emin = 23.5 -5 5

What controls the shape of the error surface What controls the shape of the error surface? Let’s examine effect of shifting the x-position of the data

Big change by simply shifting x-values of the data Region of low error is now tilted (High b, low a) has low error (Low b, high a) has low error But (high b, high a) and (low a, low b) have high error 10 5

Meaning of tilted region of low error error in (apre, bpre) are correlated

Uncorrelated estimates of intercept and slope Best-fit line Best-fit line Best fit intercept erroneous intercept When the data straddle the origin, if you tweak the intercept up, you can’t compensate by changing the slope

Negatively correlation of intercept and slope Same slope s Best-fit line Best-fit line Low slope line erroneous intercept Best fit intercept When the data are all to the right of the origin, if you tweak the intercept up, you must lower the slope to compensate

Positive correlation of intercept and slope erroneous intercept Best fit intercept Best fit intercept Same slope as best-fit line Best-fit line When the data are all to the left of the origin, if you tweak the intercept up, you must raise the slope to compensate

data near origin possibly good control on intercept but lousy control on slope small big -5 0 5

data far from origin lousy control on intercept but possibly good control on slope big 0 50 100 small

Set up for standard Least Squares yi = a + b xi y1 1 x1 a y2 = 1 x2 b … … … yN 1 xN d = G m

The formula for the least-squares solution for the general linear problem is known: mest = [GTG]-1 GT d derived by a standard minimization procedure using calculus Find the m that minimizes E(m) with E=eTe and e=d-Gm

Why least-squares. E=Si ei2 Why not least-absolute length Why least-squares? E=Si ei2 Why not least-absolute length? E=Si |ei| Or something else?

Least-Squares Least Absolute Value