Optimal Bandwidth Selection for MLS Surfaces

Slides:



Advertisements
Similar presentations
Sampling plans for linear regression
Advertisements

L1 sparse reconstruction of sharp point set surfaces
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Kriging.
An Introduction of Support Vector Machine
Data Modeling and Parameter Estimation Nov 9, 2005 PSCI 702.
Nonparametric density estimation or Smoothing the data Eric Feigelson Arcetri Observatory, April 2014.
Simple Linear Regression
Image Denoising using Locally Learned Dictionaries Priyam Chatterjee Peyman Milanfar Dept. of Electrical Engineering University of California, Santa Cruz.
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Kernel methods - overview
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Estimating Surface Normals in Noisy Point Cloud Data Niloy J. Mitra, An Nguyen Stanford University.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Lecture Notes for CMPUT 466/551 Nilanjan Ray
Shape Modeling International 2007 – University of Utah, School of Computing Robust Smooth Feature Extraction from Point Clouds Joel Daniels ¹ Linh Ha ¹.
Development of Empirical Models From Process Data
What is Learning All about ?  Get knowledge of by study, experience, or being taught  Become aware by information or from observation  Commit to memory.
1 An Introduction to Nonparametric Regression Ning Li March 15 th, 2004 Biostatistics 277.
Defining Point Set Surfaces Nina Amenta and Yong Joo Kil University of California, Davis IDAV Institute for Data Analysis and Visualization Visualization.
Robust Moving Least-squares Fitting with Sharp Features Shachar Fleishman* Daniel Cohen-Or § Claudio T. Silva* * University of Utah § Tel-Aviv university.
Classification and Prediction: Regression Analysis
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Dual Evolution for Geometric Reconstruction Huaiping Yang (FSP Project S09202) Johannes Kepler University of Linz 1 st FSP-Meeting in Graz, Nov ,
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Feature-Aware Filtering for Point-Set Surface Denoising Min Ki Park*Seung.
Gwangju Institute of Science and Technology Intelligent Design and Graphics Laboratory Multi-scale tensor voting for feature extraction from unstructured.
Using ESRI ArcGIS 9.3 Spatial Analyst
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Regression Maarten Buis Outline Recap Estimation Goodness of Fit Goodness of Fit versus Effect Size transformation of variables and effect.
Projecting points onto a point cloud with noise Speaker: Jun Chen Mar 26, 2008.
Geographic Information Science
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
GEOSTATISICAL ANALYSIS Course: Special Topics in Remote Sensing & GIS Mirza Muhammad Waqar Contact: EXT:2257.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Optimal Sampling Strategies for Multiscale Stochastic Processes Vinay Ribeiro Rolf Riedi, Rich Baraniuk (Rice University)
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Lecture 6: Point Interpolation
Chapter1: Introduction Chapter2: Overview of Supervised Learning
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Math 4030 – 11b Method of Least Squares. Model: Dependent (response) Variable Independent (control) Variable Random Error Objectives: Find (estimated)
Machine Learning 5. Parametric Methods.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Kernel Methods Arie Nakhmani. Outline Kernel Smoothers Kernel Density Estimators Kernel Density Classifiers.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Rongjie Lai University of Southern California Joint work with: Jian Liang, Alvin Wong, Hongkai Zhao 1 Geometric Understanding of Point Clouds using Laplace-Beltrami.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
The simple linear regression model and parameter estimation
Linear Regression with One Regression
Ch3: Model Building through Regression
Statistical Methods For Engineers
KERNEL DENSITY ESTIMATION
Linear regression Fitting a straight line to observations.
Nonlinear Fitting.
Parametric Methods Berlin Chen, 2005 References:
Introduction to Sensor Interpretation
Robust Moving Least-squares Fitting with Sharp Features
Introduction to Sensor Interpretation
Applied Statistics and Probability for Engineers
Probabilistic Surrogate Models
Uncertainty Propagation
Presentation transcript:

Optimal Bandwidth Selection for MLS Surfaces Hao Wang Carlos E. Scheidegger Claudio T. Silva SCI Institute – University of Utah Shape Modeling International 2008 – Stony Brook University

Point Set Surfaces Levin’s MLS formulation Shape Modeling International 2008 – Stony Brook University

Neighborhood and Bandwidth Three parameters in both steps of Levin’s MLS: Weight function Neighborhood Bandwidth Second Step of Levin’s MLS is weighted least squares polynomial fitting Bandwidth determination is important because of the overfitting/underfitting problem Overfitting Underfitting Shape Modeling International 2008 – Stony Brook University

Neighborhood and Bandwidth Common practice Weight function: Exponential Neighborhood: Spherical Bandwidth: Heuristics Problems Optimality Anisotropic Dataset Weight function: Exponential function is computationally intensive. Other function may lead to higher computational efficiency or better reconstruction quality. The APSS paper (SIGGRAPH 2007) uses (1-x^2)^4 as the weight function. Neighborhood: Spherical neighborhood may not deal with anisotropic data properly. Other neighborhood shape may lead to better reconstruction quality. Bandwidth: Empirically choose bandwidth based on trial and error involves human interaction. In addition, even though heuristic methods can produce visually acceptable reconstruction, the result might not be geometrically accurate. Shape Modeling International 2008 – Stony Brook University

Related Work Other MLS Formulations Alexa et al. Guennebaud et al. Robust Feature Extraction Fleishman et al. Bandwidth Determination Adamson et al. Lipman et al. Guennebaud et al. “APSS” SIGGRAPH 2007

Locally Weighted Kernel Regression Problem Points sampled from functional with white noise added White noise are i.i.d. random variables Reconstruct the functional with least squares criterion Approach Consider each point p individually p is reconstructed by utilizing information of its neighborhood Influence of each neighboring point is related to its distance from p Shape Modeling International 2008 – Stony Brook University

Kernel Regression v.s MLS Surfaces Kernel Regression is mostly the same as the second step in Levin’s MLS. The only difference is between kernel weighting and MLS weighting. Shape Modeling International 2008 – Stony Brook University

Kernel Regression v.s MLS Surfaces Difference Kernel weighting for functional data MLS weighting for manifold data Advantages of Kernel Regression More mature technique for processing noisy sample points Behavior of the neighborhood and kernel better studied Goal Adapt techniques in kernel regression to MLS surfaces Extend theoretical results of kernel regression to MLS surfaces Shape Modeling International 2008 – Stony Brook University

Weight Function Common choices of weight functions in kernel regression: Epanechnikov Normal Biweight Optimal weight function: Epanechnikov Choice of weight function not important Implication: Optimality Most weight functions produce results with ignorable differences Implications: Exponential weight function in Levin’s MLS can be replaced by weight functions requiring less computational effort Weight function in Levin’s MLS is indeed a near optimal choice Shape Modeling International 2008 – Stony Brook University

Evaluation of Kernel Regression MSE MSE = Mean Squared Error Evaluate result of the functional fitting at each point Shape Modeling International 2008 – Stony Brook University

Evaluation of Kernel Regression MISE Integration of MSE over the domain Evaluate the global performance of kernel regression Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth Optimality Computation Leading to minimum MSE / MISE Each point with a different optimal bandwidth Computation MSE / MISE approximated by Taylor Polynomial Solve for the minimizing bandwidth Computation MSE / MISE can be approximated by a Taylor polynomial The Taylor polynomial involves bandwidth and thus can be considered as a function of bandwidth Solve analytically for minimizing bandwidth of the polynomial by setting its derivative to be 0 and solve the equation Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth Approach Unknown quantities in computation Derivatives of underlying functional Variance of random noise variables Density of point set Approach Derivatives: Ordinary Least Squares Fitting Variance: Statistical Inference Density: Kernel Density Estimation Computation of bandwidth is difficult because it involves unknown quantities: Taylor polynomial involves derivatives of underlying functional Taylor polynomial involves variance of random noise variables Taylor polynomial may involve density of point set Solution Underlying functional approximated by ordinary least squares fitting Derivatives approximated by derivatives of approximated functional Variance of random noise variables estimated by statistical inference Density of point set estimated by kernel density estimation Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth in 2-D Optimal bandwidth based on MSE: Interpretation Higher noise level : larger bandwidth Higher curvature : smaller bandwidth Higher density : smaller bandwidth More point samples : smaller bandwidth Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth in 3-D Kernel Function: with Kernel Shape: Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth in 3-D Optimal spherical bandwidth based on MSE: Optimal spherical bandwidth based on MISE: Shape Modeling International 2008 – Stony Brook University

Experiments Bandwidth selectors choose near optimal bandwidths Shape Modeling International 2008 – Stony Brook University

Experiments Shape Modeling International 2008 – Stony Brook University

Experiments Shape Modeling International 2008 – Stony Brook University

Optimal Bandwidth in MLS From functional domain to manifold domain Choose a functional domain Use kernel regression with modification There are 2 ways to do this: 1. Choose functional domain using k-NN or constant radius sphere and apply kernel regression with optimal bandwidth 2. Use kernel regression with modification Approximate underlying function using MLS with heuristically chosen bandwidths Approximate unknown quantities in optimal bandwidth formulas using approximated MLS surface Shape Modeling International 2008 – Stony Brook University

Robustness Insensitivity to error in first step of Levin’s MLS Reference plane found by the first step is rotated by angle theta, as shown on x-axis. The y-axis shows the mean value of the bandwidth / reconstruction error. The angle v.s bandwidth plot shows a curve similar to plot of cosine function because by rotating theta, roughly cos(theta) of original point set remains. Shape Modeling International 2008 – Stony Brook University

Comparison Constant h: uniform v.s non-uniform sampling k-NN: sampling v.s feature MSE/MISE based plug-in method: most robust and flexible Constant h: Not suitable for non-uniformly point samples because of its lack of local adaptation. k-NN: Can not distinguish high noise level and high curvature MSE/MISE based plug-in method: Work for both regularly and irregularly sampled points Capable of distinguishing noise and curvature Shape Modeling International 2008 – Stony Brook University

Comparison MSE/MISE-based plug-in method better than heuristic methods Shape Modeling International 2008 – Stony Brook University

Comparison Heuristic methods can produce visually acceptable but not geometrically accurate reconstruction. Shape Modeling International 2008 – Stony Brook University

Future Work Nonlinear kernel regression bandwidth selector in 3-D Compute optimal bandwidth implicitly Extend the method to other MLS formulations Nonlinear kernel regression bandwidth selector in 3-D Derivation mathematically intense Formulas need to be simplified Compute optimal bandwidth implicitly Avoiding computing bandwidth analytically may lead to do simpler method Extend the method to other MLS formulations Shape Modeling International 2008 – Stony Brook University