Geo597 Geostatistics Ch11 Point Estimation. Point Estimation  In the last chapter, we looked at estimating a mean value over a large area within which.

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Copyright © 2010 Pearson Education, Inc. Slide
Sampling: Final and Initial Sample Size Determination
Simple Linear Regression. G. Baker, Department of Statistics University of South Carolina; Slide 2 Relationship Between Two Quantitative Variables If.
Objectives (BPS chapter 24)
AP Statistics Chapters 3 & 4 Measuring Relationships Between 2 Variables.
Geo479/579: Geostatistics Ch14. Search Strategies.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Basic Statistical Concepts Psych 231: Research Methods in Psychology.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 171 CURVE.
Slide Copyright © 2010 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Business Statistics First Edition.
Lecture 19 Transformations, Predictions after Transformations Other diagnostic tools: Residual plot for nonconstant variance, histogram to check normality.
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
Business Statistics - QBM117 Statistical inference for regression.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Relationships Among Variables
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Correlation & Regression
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Introduction to Linear Regression and Correlation Analysis
Inference for regression - Simple linear regression
Geo479/579: Geostatistics Ch13. Block Kriging. Block Estimate  Requirements An estimate of the average value of a variable within a prescribed local.
STA291 Statistical Methods Lecture 27. Inference for Regression.
CORRELATION & REGRESSION
Geo479/579: Geostatistics Ch17. Cokriging
Inferences for Regression
BPS - 3rd Ed. Chapter 211 Inference for Regression.
Geo479/579: Geostatistics Ch12. Ordinary Kriging (1)
Measures of Variability In addition to knowing where the center of the distribution is, it is often helpful to know the degree to which individual values.
Chapter 4 Variability. Variability In statistics, our goal is to measure the amount of variability for a particular set of scores, a distribution. In.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Model Construction: interpolation techniques 1392.
Examining Relationships in Quantitative Research
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
Chapter 14 Inference for Regression AP Statistics 14.1 – Inference about the Model 14.2 – Predictions and Conditions.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Regression Regression relationship = trend + scatter
Measures of Dispersion
Spatial Interpolation Chapter 13. Introduction Land surface in Chapter 13 Land surface in Chapter 13 Also a non-existing surface, but visualized as a.
Geo479/579: Geostatistics Ch4. Spatial Description.
Lecture 10 Chapter 23. Inference for regression. Objectives (PSLS Chapter 23) Inference for regression (NHST Regression Inference Award)[B level award]
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
Geo479/579: Geostatistics Ch15. Cross Validation.
Robust Estimators.
AP Statistics Semester One Review Part 1 Chapters 1-3 Semester One Review Part 1 Chapters 1-3.
Chapter 14: Inference for Regression. A brief review of chapter 4... (Regression Analysis: Exploring Association BetweenVariables )  Bi-variate data.
2010.  A first degree equation is called a linear equation in two variables if it contains two distinct variables.
Review Normal Distributions –Draw a picture. –Convert to standard normal (if necessary) –Use the binomial tables to look up the value. –In the case of.
Introduction to statistics I Sophia King Rm. P24 HWB
Geo479/579: Geostatistics Ch12. Ordinary Kriging (2)
Geo479/579: Geostatistics Ch10. Global Estimation.
Univariate Point Estimation Confidence Interval Estimation Bivariate: Linear Regression Multivariate: Multiple Regression 1 Chapter 4: Statistical Approaches.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Chapter 4 Variability PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Seventh Edition by Frederick J Gravetter and Larry.
Chapter 4 More on Two-Variable Data. Four Corners Play a game of four corners, selecting the corner each time by rolling a die Collect the data in a table.
Part 5 - Chapter
Part 5 - Chapter 17.
Correlation, Bivariate Regression, and Multiple Regression
AP Statistics Chapter 14 Section 1.
Basic Estimation Techniques
Ch9 Random Function Models (II)
Basic Estimation Techniques
CHAPTER 29: Multiple Regression*
Part 5 - Chapter 17.
Lecture 20: Spatial Interpolation III
Basic Practice of Statistics - 3rd Edition Inference for Regression
Inferences for Regression
Presentation transcript:

Geo597 Geostatistics Ch11 Point Estimation

Point Estimation  In the last chapter, we looked at estimating a mean value over a large area within which there are many samples.  Eventually we need to estimate unknown values at specific locations, using weighted linear combinations.  In addition to clustering, we have to account for the distance to the nearby samples.

In This Chapter  Four methods for point estimation, polygons, triangulation, local sample means, and inverse distance.  Statistical tools to evaluate the performance of these methods.

Polygon  Same as the polygonal declustering method for global estimation.  The value of the closest sample point is simply chosen as the estimate of the point of interest.  It can be viewed as a weighted linear combination with all the weights given to a single sample, the closest one.

Polygon...  As long as the point of interest falls within the same polygon of influence, the polygonal estimate remains the same

?=696

Triangulation  Discontinuities in the polygonal estimation are often unrealistic.  Triangulation methods remove the discontinuities by fitting a plane through three samples that surround the point being estimated.

Triangulation...  Equation of the plane: (z is the V value, x is the easting, and y is the northing)  Given the coordinates and V value of the 3 nearby samples, coefficients a, b, and c can be calculated by solving the following system equations:

Triangulation... 63a + 140b + c = a + 129b + c = a + 140b + c = 606 a = , b = , c = = x y  This is the equation of the plane passing through the three nearby samples.  We can now estimate the value of any location in the plane as long as we have the x, y, and z.

?=548.7 = * *

Triangulation...  Triangulation estimate depends on which three nearby sample points are chosen to form a plane.  Delaunay triangulation, a particular triangulation, produces triangles that are as close to equilateral as possible.  Three sample locations form a Delaunay triangle if their polygons of influence share a common vertex.

Triangulation...  Triangulation is not used for extrapolation beyond the edges of the triangle.  Triangulation estimate can also be expressed as a weighted linear combination of the three sample values.  Each sample value is weighted according to the area of the opposite triangle.

?=548.7=[(22.5)(696)+(12)(227)+(9.5)(606)]/44

Local Sample Mean  This method weights all nearby samples equally, and uses the sample mean as the estimate. It is a weighted linear combination of equal weights.  This is the first step in the cell declustering in ch10.  This approach is spatially naïve.

?=603.7=( )/7

Inverse Distance Methods  Weight each sample inversely proportional to any power of its distance from the point being estimated:  It is obviously a weighted linear combination

ID SAMP#XYVDist1/d i (1/d i )/( 1/d i ) /d i = Table 11.2 Mean is 603.7

# V p=0.2 p=0.5 p=1.0 p=2.0 p=5.0 p= < < < <.0001 V p = exponent Local sample mean = Polygonal estimate = 696 Table 11.3

?=594 (p=1) =477* * * * * * *0.07

Inverse Distance Methods...  As p approaches 0, the weights become more similar and the estimate approaches the simple local sample mean,d 0 =1.  As p approaches, the estimate approaches the polygonal estimate, giving all of the weight to the closest sample.

Estimation Criteria  Best and unbiased  MAE and MSE  Global and conditional unbiased  Smoothing effect

Estimation Criteria  Univariate Distribution of Estimates  The distribution of estimated values should be close to that of the true values.  Compare the mean, medians, and standard deviation between the estimated and the true.  The q-q plot of the estimated and the true distributions often reveal subtle differences that are hard to detect with only a few summary statistics.

Estimation Criteria...  Univariate Distribution of Errors  Error (residuals) =  Preferable conditions of the error distribution 1. Unbiased estimate the mean of the error distribution is referred to as bias unbiased: Median(r) = 0; mode(r) = 0 (balanced over- and under-estimates, and symmetric error distribution).

Estimation Criteria...  Univariate Distribution of Errors...  Preferable conditions of the error distribution 2. Small spread Small standard deviation or variance of errors  A small spread is preferred to a small bias (remember the proportional effect?)

Less variability is preferred to a small bias Remember a similar concept when we discussed something similar in proportional effect?

Estimation Criteria...  Summary statistics of bias and spread - Mean Absolute Error (MAE) = - Mean Squared Error (MSE) =

Estimation Criteria  Ideally, it is desirable to have unbiased distribution for each of the many subgroups of estimates (conditional unbiasedness, Fig 3.6, p36).  A set of estimates that is conditionally unbiased is also globally unbiased, however the reverse is not true.  One way of checking for conditional bias is to plot the errors against the estimated values.

Conditional Unbiasedness

Estimation Criteria...  Bivariate Distribution of Estimated and True Values  Scatter plot of true versus predicted values.  The best possible estimates would always match the true values and would therefore plot on the 45-degree line on a scatterplot.

Estimation Criteria...  Bivariate Distribution of Estimated and True Values...  If the mean error is zero for any range of estimated values, the conditional expectation curve of true values given estimated ones will plot on the 45-degree line.

Case Studies  Different estimation methods have different smoothing effects (reduced variability of estimated values).  The more sample points are used for an estimation, the smoother the estimate would become (ch14).  The polygonal method uses only one sample, thus un-smoothed.  Smoothed estimates contain fewer extreme values.

Distribution of estimated vs. true values

Effect of clustered data on global estimates

Which is the best?  We like to have a method that uses the nearby samples and also accounts for the clustering in the samples configuration

Detecting Conditional Biasedness