Basis Expansion and Regularization

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Lecture 4. Linear Models for Regression
Flexible smoothing with B-splines and Penalties or P-splines P-splines = B-splines + Penalization Applications : Generalized Linear and non linear Modelling.
Pattern Recognition and Machine Learning
Pattern Recognition and Machine Learning: Kernel Methods.
Model assessment and cross-validation - overview
Data mining and statistical learning - lecture 6
Selected from presentations by Jim Ramsay, McGill University, Hongliang Fei, and Brian Quanz Basis Basics.
Basis Expansion and Regularization Presenter: Hongliang Fei Brian Quanz Brian Quanz Date: July 03, 2008.
1 Chapter 4 Interpolation and Approximation Lagrange Interpolation The basic interpolation problem can be posed in one of two ways: The basic interpolation.
Vector Generalized Additive Models and applications to extreme value analysis Olivier Mestre (1,2) (1) Météo-France, Ecole Nationale de la Météorologie,
Data mining in 1D: curve fitting
Jim Ramsay McGill University Basis Basics. Overview  What are basis functions?  What properties should they have?  How are they usually constructed?
Support Vector Machine
Pattern Recognition and Machine Learning
Kernel methods - overview
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
統計計算與模擬 政治大學統計系余清祥 2003 年 6 月 9 日 ~ 6 月 10 日 第十六週:估計密度函數
x – independent variable (input)
Some useful linear algebra. Linearly independent vectors span(V): span of vector space V is all linear combinations of vectors v i, i.e.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Linear Regression Models Based on Chapter 3 of Hastie, Tibshirani and Friedman Slides by David Madigan.
1 Introduction to Kernels Max Welling October (chapters 1,2,3,4)
NOTES ON MULTIPLE REGRESSION USING MATRICES  Multiple Regression Tony E. Smith ESE 502: Spatial Data Analysis  Matrix Formulation of Regression  Applications.
Basis Expansions and Regularization Based on Chapter 5 of Hastie, Tibshirani and Friedman.
Basic Concepts and Definitions Vector and Function Space. A finite or an infinite dimensional linear vector/function space described with set of non-unique.
Nonparametric Smoothing Methods and Model Selections T.C. Lin Dept. of Statistics National Taipei University 5/4/2005.
Kernel Methods and SVM’s. Predictive Modeling Goal: learn a mapping: y = f(x;  ) Need: 1. A model structure 2. A score function 3. An optimization strategy.
Linear and generalised linear models
1 An Introduction to Nonparametric Regression Ning Li March 15 th, 2004 Biostatistics 277.
Subdivision Analysis via JSR We already know the z-transform formulation of schemes: To check if the scheme generates a continuous limit curve ( the scheme.
(CH. 5 Part 2) Speaker: Brian Quanz 7/3/2008
Outline Separating Hyperplanes – Separable Case
V. Space Curves Types of curves Explicit Implicit Parametric.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
Basis Expansions and Regularization Part II. Outline Review of Splines Wavelet Smoothing Reproducing Kernel Hilbert Spaces.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Sparse Kernel Methods 1 Sparse Kernel Methods for Classification and Regression October 17, 2007 Kyungchul Park SKKU.
Linear Models for Classification
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Kernel Methods Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
1. Systems of Linear Equations and Matrices (8 Lectures) 1.1 Introduction to Systems of Linear Equations 1.2 Gaussian Elimination 1.3 Matrices and Matrix.
Additive Models , Trees , and Related Models Prof. Liqing Zhang Dept. Computer Science & Engineering, Shanghai Jiaotong University.
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Linear Classifiers Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
1 Chapter 4 Interpolation and Approximation Lagrange Interpolation The basic interpolation problem can be posed in one of two ways: The basic interpolation.
1 Kernel-class Jan Recap: Feature Spaces non-linear mapping to F 1. high-D space 2. infinite-D countable space : 3. function space (Hilbert.
Basis Expansions and Generalized Additive Models Basis expansion Piecewise polynomials Splines Generalized Additive Model MARS.
LECTURE 17: BEYOND LINEARITY PT. 2 March 30, 2016 SDS 293 Machine Learning.
Model Selection and the Bias–Variance Tradeoff All models described have a smoothing or complexity parameter that has to be considered: multiplier of the.
CS Statistical Machine learning Lecture 7 Yuan (Alan) Qi Purdue CS Sept Acknowledgement: Sargur Srihari’s slides.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Piecewise Polynomials and Splines
ECE3340 Numerical Fitting, Interpolation and Approximation
Dept. Computer Science & Engineering, Shanghai Jiao Tong University
Dept. Computer Science & Engineering, Shanghai Jiao Tong University
Machine learning, pattern recognition and statistical data modelling
Machine learning, pattern recognition and statistical data modelling
Bias and Variance of the Estimator
Biointelligence Laboratory, Seoul National University
Basis Expansions and Generalized Additive Models (2)
Basis Expansions and Generalized Additive Models (1)
政治大學統計系余清祥 2004年5月26日~ 6月7日 第十六、十七週:估計密度函數
Support Vector Machines 2
Presentation transcript:

Basis Expansion and Regularization Prof. Liqing Zhang Dept. Computer Science & Engineering, Shanghai Jiaotong University

Basis Expansion and Regularization Outline Piece-wise Polynomials and Splines Wavelet Smoothing Smoothing Splines Automatic Selection of the Smoothing Parameters Nonparametric Logistic Regression Multidimensional Splines Regularization and Reproducing Kernel Hilbert Spaces 2017/4/15 Basis Expansion and Regularization

Piece-wise Polynomials and Splines Linear basis expansion Some basis functions that are widely used 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Three approaches for controlling the complexity of the model. Restriction Selection Regularization: 2017/4/15 Basis Expansion and Regularization

Piecewise Polynomials and Splines 2017/4/15 Basis Expansion and Regularization

Piecewise Cubic Polynomials Increasing orders of continuity at the knots. A cubic spline with knots at and : Cubic spline truncated power basis 2017/4/15 Basis Expansion and Regularization

Piecewise Cubic Polynomials An order-M spline with knots , j=1,…,K is a piecewise-polynomial of order M, and has continuous derivatives up to order M-2. A cubic spline has M=4. Truncated power basis set: 2017/4/15 Basis Expansion and Regularization

Natural cubic spline Natural cubic spline adds additional constraints, namely that the function is linear beyond the boundary knots. Linear Cubic Natural boundary constraints 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization B-spline The augmented knot sequenceτ: Bi,m(x), the i-th B-spline basis function of order m for the knot-sequenceτ, m≤M. 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization B-spline The sequence of B-spline up to order 4 with ten knots evenly spaced from 0 to 1 The B-spline have local support; they are nonzero on an interval spanned by M+1 knots. 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing Splines Base on the spline basis method: So , is the noise. Minimize the penalized residual sum of squares is a fixed smoothing parameter f can be any function that interpolates the data the simple least squares line fit 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing Splines The solution is a natural spline: Then the criterion reduces to: where So the solution: The fitted smoothing spline: 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing Splines Function of age, that response the relative change in bone mineral density measured at the spline in adolescents Separate smoothing splines fit the males and females, 12 degrees of freedom 脊骨BMD—骨质密度 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing Matrix the N-vector of fitted values The finite linear operator — the smoother matrix Compare with the linear operator in the LS-fitting: M cubic-spline basis functions, knot sequenceξ Similarities and differences: Both are symmetric, positive semidefinite matrices idempotent(幂等的) ; shrinking rank: 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing Matrix Effective degrees of freedom of a smoothing spline in the Reinsch form: Since , solution: is symmetric and has a real eigen-decomposition is the corresponding eigenvalue of K 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Smoothing spline fit of ozone(臭氧) concentration versus Daggot pressure gradient. Smoothing parameter df=5 and df=10. The 3rd to 6th eigenvectors of the spline smoothing matrices 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization The smoother matrix for a smoothing spline is nearly banded, indicating an equivalent kernel with local support. 2017/4/15 Basis Expansion and Regularization

Bias-Variance Tradeoff Example: For The diagonal contains the pointwise variances at the training Bias is given by is the (unknown) vector of evaluations of the true f 2017/4/15 Basis Expansion and Regularization

Bias-Variance Tradeoff df=5, bias high, standard error band narrow df=9, bias slight, variance not increased appreciably df=15, over learning, standard error widen 2017/4/15 Basis Expansion and Regularization

Bias-Variance Tradeoff The EPE and CV curves have the a similar shape. And, overall the CV curve is approximately unbiased as an estimate of the EPE curve The integrated squared prediction error (EPE) combines both bias and variance in a single summary: N fold (leave one) cross-validation: 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Logistic Regression Logistic regression with a single quantitative input X The penalized log-likelihood criterion 2017/4/15 Basis Expansion and Regularization

Multidimensional Splines Tensor product basis The M1×M2 dimensional tensor product basis , basis function for coordinate X1 , basis function for coordinate X2 2017/4/15 Basis Expansion and Regularization

Tenor product basis of B-splines, some selected pairs 2017/4/15 Basis Expansion and Regularization

Multidimensional Splines High dimension smoothing Splines J is an appropriate penalty function a smooth two-dimensional surface, a thin-plate spline. The solution has the form 2017/4/15 Basis Expansion and Regularization

Multidimensional Splines The decision boundary of an additive logistic regression model. Using natural splines in each of two coordinates. df = 1 +(4-1) + (4-1) = 7 2017/4/15 Basis Expansion and Regularization

Multidimensional Splines The results of using a tensor product of natural spline basis in each coordinate. df = 4 x 4 = 16 2017/4/15 Basis Expansion and Regularization

Multidimensional Splines A thin-plate spline fit to the heart disease data. The data points are indicated, as well as the lattice of points used as knots. 2017/4/15 Basis Expansion and Regularization

Reproducing Kernel Hilbert space A regularization problems has the form: L(y,f(x)) is a loss-function. J(f) is a penalty functional, and H is a space of functions on which J(f) is defined. The solution span the null space of the penalty functional J 2017/4/15 Basis Expansion and Regularization

Spaces of Functions Generated by Kernel Important subclass are generated by the positive kernel K(x,y). The corresponding space of functions Hk is called reproducing kernel Hilbert space. Suppose that K has an eigen-expansion Elements of H have an expansion 2017/4/15 Basis Expansion and Regularization

Spaces of Functions Generated by Kernel The regularization problem become The finite-dimension solution(Wahba,1990) Reproducing properties of kernel function 2017/4/15 Basis Expansion and Regularization

Spaces of Functions Generated by Kernel The penalty functional The regularization function reduces to a finite-dimensional criterion K is NxN matrix, the ij-th entry K(xi, xj) 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization RKHS Penalized least squares The solution of : The fitted values: The vector of N fitted value is given by 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Example of RKHS Polynomial regression Suppose M huge Given Loss function: The penalty polynomial regression: 2017/4/15 Basis Expansion and Regularization

Penalized Polynomial Regression Kernel: has eigen-functions E.g. d=2, p=2: The penalty polynomial regression: 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization RBF kernel & SVM kernel Gaussian Radial Basis Functions Support Vector Machines 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Wavelet smoothing Another type of bases——Wavelet bases Wavelet bases are generated by translations and dilations of a single scaling function . If ,then generates an orthonormal basis for functions with jumps at the integers. form a space called reference space V0 The dilations form an orthonormal basis for a space V1 V0 Generally, we have 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Wavelet smoothing The L2 space dividing Mother wavelet generate function form an orthonormal basis for W0. Likewise form a basis for Wj. 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization Wavelet smoothing Wavelet basis on W0 : Wavelet basis on Wj : The symmlet-p wavelet: A support of 2p-1 consecutive intervals. p vanishing moments: 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization

Adaptive Wavelet Filtering Wavelet transform: y: response vector, W: NxN orthonormal wavelet basis matrix Stein Unbiased Risk Estimation (SURE) The solution: Fitted function is given by inverse wavelet transform: , LS coefficients truncated to 0 Simple choice for 2017/4/15 Basis Expansion and Regularization

Basis Expansion and Regularization 2017/4/15 Basis Expansion and Regularization