USE OF KERNELS FOR HYPERSPECTRAL TRAGET DETECTION Nasser M. Nasrabadi Senior Research Scientist U.S. Army Research Laboratory, Attn: AMSRL-SE-SE 2800 Powder.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Outlines Background & motivation Algorithms overview
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Olivier Duchenne , Armand Joulin , Jean Ponce Willow Lab , ICCV2011.
1 Welcome to the Kernel-Class My name: Max (Welling) Book: There will be class-notes/slides. Homework: reading material, some exercises, some MATLAB implementations.
Input Space versus Feature Space in Kernel- Based Methods Scholkopf, Mika, Burges, Knirsch, Muller, Ratsch, Smola presented by: Joe Drish Department of.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Pattern Recognition and Machine Learning: Kernel Methods.
CS Statistical Machine learning Lecture 13 Yuan (Alan) Qi Purdue CS Oct
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Pattern Recognition and Machine Learning
Principal Component Analysis
Reduced Support Vector Machine
1 Introduction to Kernels Max Welling October (chapters 1,2,3,4)
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Continuous Latent Variables --Bishop
Comparing Kernel-based Learning Methods for Face Recognition Zhiguo Li
Learning in Feature Space (Could Simplify the Classification Task)  Learning in a high dimensional space could degrade generalization performance  This.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Unsupervised learning
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
A New Subspace Approach for Supervised Hyperspectral Image Classification Jun Li 1,2, José M. Bioucas-Dias 2 and Antonio Plaza 1 1 Hyperspectral Computing.
Chapter 21 R(x) Algorithm a) Anomaly Detection b) Matched Filter.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Part 2: Change detection for SAR Imagery (based on Chapter 18 of the Book. Change-detection methods for location of mines in SAR imagery, by Dr. Nasser.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Texture scale and image segmentation using wavelet filters Stability of the features Through the study of stability of the eigenvectors and the eigenvalues.
Baseband Demodulation/Detection
Kernel adaptive filtering Lecture slides for EEL6502 Spring 2011 Sohan Seth.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
Chapter 12 Object Recognition Chapter 12 Object Recognition 12.1 Patterns and pattern classes Definition of a pattern class:a family of patterns that share.
The Restricted Matched Filter for Distributed Detection Charles Sestok and Alan Oppenheim MIT DARPA SensIT PI Meeting Jan. 16, 2002.
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Dd Generalized Optimal Kernel-based Ensemble Learning for HS Classification Problems Generalized Optimal Kernel-based Ensemble Learning for HS Classification.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Machine Learning CUNY Graduate Center Lecture 6: Linear Regression II.
Irena Váňová. B A1A1. A2A2. A3A3. repeat until no sample is misclassified … labels of classes Perceptron algorithm for i=1...N if then end * * * * *
Performance of Digital Communications System
Unsupervised Learning II Feature Extraction
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Spectral Methods for Dimensionality
PATTERN COMPARISON TECHNIQUES
Ch 12. Continuous Latent Variables ~ 12
Probability Theory and Parameter Estimation I
Background on Classification
Orthogonal Subspace Projection - Matched Filter
CHE 391 T. F. Edgar Spring 2012.
PCA vs ICA vs LDA.
IX International Workshop ACAT
Principal Component Analysis
EE513 Audio Signals and Systems
Biointelligence Laboratory, Seoul National University
Digital Image Processing Week IV
Generally Discriminant Analysis
Parametric Methods Berlin Chen, 2005 References:
APPENDIX B Multivariate Statistics
16. Mean Square Estimation
NOISE FILTER AND PC FILTERING
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

USE OF KERNELS FOR HYPERSPECTRAL TRAGET DETECTION Nasser M. Nasrabadi Senior Research Scientist U.S. Army Research Laboratory, Attn: AMSRL-SE-SE 2800 Powder Mill Road, Adelphi, MD 20783, USA

2 Outline Develop nonlinear detection algorithms. Exploiting higher order correlations. Why Kernels Kernel Trick Conventional matched filters Kernel matched filters Detection results

Nonlinear Mapping of Data Exploitation of Nonlinear Correlations Nonlinear mapping Statistical learning (VC): Mapping into a higher dimensional space increases data separability However, because of the infinite dimensionality implementing conventional detectors in the feature space is not feasible using conventional methods Convert the detector expression into dot product forms Kernel-based nonlinear version of the conventional detector Input space High dimensional feature space Input space Kernel trick :

4 Kernel Trick Example of the kernel trick Consider 2-D input patterns, where If a 2nd order monomial is used as the nonlinear mapping This property generalizes for and

5 Examples of Kernels 1. Gaussian RBF kernel: 4. Polynomial kernel: 3. Spectral angle-based kernel: 2. Inverse multiquadric kernel: Possible realization of

Linear Spectral Matched Filter Spectral signal model target present target spectral signature, background clutter noise Linear matched filter design (signal-to-clutter ratio)  - Average output power of the filter for - Constrained energy minimization: no target, target present

To stabilize the inverse of the covariance matrix, usually regularization is used, equivalent to minimizing: The regularized matched filter is given by: Linear Spectral Matched Filter & Regularized Spectral Matched Filter Linear matched filter is given as:

8 Nonlinear Spectral Matched Filter In the feature space, the equivalent signal model The equivalent matched filter in the feature space Output of the matched filter in the feature space

Kernelization of Spectral Matched Filter in Feature space Using the following properties of PCA and Kernel PCA Each eigenvector can be represented in terms of the input data Inverse Covariance matrix is now Kernel matrix spectral decomposition (kernel PCA) The kernelized version of matched filter

10 Conventional MF vs. Kernel MF Conventional spectral matched filter Nonlinear matched filter Kernel matched filter

11 Matched Subspace Detection (MSD) Consider a linear mixed model: where and represent orthogonal matrices whose column vectors span the target and background subspaces and are unknown vectors of coefficients, is a Gaussian random noise distributed as The log Generalized likelihood ratio test (GLRT) is given by where

12 Kernel Matched Subspace Detection Define the matched subspace detector in the feature space To kernelize we use the kernel PCA, and kernel function properties as shown below

13 MSD vs. Kernel MSD GLRT for the MSD: Kernelized GLRT for the kernel MSD: Nonlinear GLRT for the MSD in feature space:

14 The model in the nonlinear feature space is The MLE for   in feature space is given as The kernel version of   is given as Orthogonal Subspace Projector vs. Kernel OSP

15 Consider a linear mixed model: where U represent an orthogonal matrix whose column vectors span the target subspace and C is the background covariance. is unknown vector of coefficients, is a Gaussian random noise distributed as The log Generalized likelihood ratio test (GLRT) is given by Adaptive Subspace Detection (ASD)

16 The model in the nonlinear feature space is The GLRT ASD in feature space is given as Where  is a nonlinear function. Substituting the following identities into the above Eq. Nonlinear ASD

17 ASD vs. Kernel ASD GLRT for the ASD: Kernelized GLRT for the kernel ASD: Nonlinear GLRT for the ASD in feature space:

18 A 2-D Gaussian Toy Example Red dots belong to class H 1, blue dots belong to H 0 (a)MSD (h) KSMF(b) KMSD (c) ASD (d) KASD (e) OSP (f) KOSP (g) SMF

19 Red dots belong to class H 1, blue dots belong to H 0 A 2-D Toy Example (a)MSD (h) KSMF(b) KMSD (c) ASD (d) KASD (e) OSP (f) KOSP (g) SMF

20 Test Images Forest Radiance I Desert Radiance II

21 Results for DR-II Image (a)MSD (h) KSMF(b) KMSD (c) ASD (d) KASD (e) OSP (f) KOSP (g) SMF

22 ROC Curves for DR-II Image

23 (a)MSD (h) KSMF(b) KMSD (c) ASD (d) KASD (e) OSP (f) KOSP (g) SMF Results for FR-II Image

24 ROC Curves for FR-II Image

25 SMF & KSMF Results for Mine Image Mine Hyperspectral Image KSMF for mine imageSMF for mine image

26 ROC Curves for Mine Image

27 Conclusions Nonlinear target detection techniques are valuable. Use of kernels and regularization in filter design. Choice of kernels? Nonlinear sensor fusion using kernels.