An M-ary KMP Classifier for Multi-aspect Target Classification

Slides:



Advertisements
Similar presentations
Active Shape Models Suppose we have a statistical shape model –Trained from sets of examples How do we use it to interpret new images? Use an “Active Shape.
Advertisements

STOR 892 Object Oriented Data Analysis Radial Distance Weighted Discrimination Jie Xiong Advised by Prof. J.S. Marron Department of Statistics and Operations.
Introduction to Neural Networks Computing
On-line learning and Boosting
Multi-Label Prediction via Compressed Sensing By Daniel Hsu, Sham M. Kakade, John Langford, Tong Zhang (NIPS 2009) Presented by: Lingbo Li ECE, Duke University.
Pattern Recognition and Machine Learning
Computer vision: models, learning and inference Chapter 8 Regression.
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Optimal Adaptation for Statistical Classifiers Xiao Li.
Radial-Basis Function Networks
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Mining Discriminative Components With Low-Rank and Sparsity Constraints for Face Recognition Qiang Zhang, Baoxin Li Computer Science and Engineering Arizona.
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering National Institute of Technology.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
1 Learning with Local and Global Consistency Presented by Qiuhua Liu Duke University Machine Learning Group March 23, 2007 By Dengyong Zhou, Olivier Bousquet,
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
ECE738 Advanced Image Processing Face Detection IEEE Trans. PAMI, July 1997.
Non-Bayes classifiers. Linear discriminants, neural networks.
12/4/981 Automatic Target Recognition with Support Vector Machines Qun Zhao, Jose Principe Computational Neuro-Engineering Laboratory Department of Electrical.
Biointelligence Laboratory, Seoul National University
SAR-ATR-MSTAR TARGET RECOGNITION FOR MULTI-ASPECT SAR IMAGES WITH FUSION STRATEGIES ASWIN KUMAR GUTTA.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Iterative similarity based adaptation technique for Cross Domain text classification Under: Prof. Amitabha Mukherjee By: Narendra Roy Roll no: Group:
Learning Kernel Classifiers Chap. 3.3 Relevance Vector Machine Chap. 3.4 Bayes Point Machines Summarized by Sang Kyun Lee 13 th May, 2005.
Seungchan Lee Department of Electrical and Computer Engineering Mississippi State University RVM Implementation Progress.
Multi-label Prediction via Sparse Infinite CCA Piyush Rai and Hal Daume III NIPS 2009 Presented by Lingbo Li ECE, Duke University July 16th, 2010 Note:
Nonlinear Adaptive Kernel Methods Dec. 1, 2009 Anthony Kuh Chaopin Zhu Nate Kowahl.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
DISPLACED PHASE CENTER ANTENNA SAR IMAGING BASED ON COMPRESSED SENSING Yueguan Lin 1,2,3, Bingchen Zhang 1,2, Wen Hong 1,2 and Yirong Wu 1,2 1 National.
Predictive Automatic Relevance Determination by Expectation Propagation Y. Qi T.P. Minka R.W. Picard Z. Ghahramani.
Neural networks and support vector machines
Semi-Supervised Clustering
Chapter 7. Classification and Prediction
CEE 6410 Water Resources Systems Analysis
Multiplicative updates for L1-regularized regression
Results for all features Results for the reduced set of features
Rule Induction for Classification Using
Bag-of-Visual-Words Based Feature Extraction
Empirical risk minimization
Session 7: Face Detection (cont.)
CSE 4705 Artificial Intelligence
Supervised Time Series Pattern Discovery through Local Importance
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Computer vision: models, learning and inference
An Enhanced Support Vector Machine Model for Intrusion Detection
Overview of Supervised Learning
RECURRENT NEURAL NETWORKS FOR VOICE ACTIVITY DETECTION
Classification / Regression Neural Networks 2
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
Boosting Nearest-Neighbor Classifier for Character Recognition
Tomasz Maszczyk and Włodzisław Duch Department of Informatics,
Lecture 24 Radial Basis Network (I)
Neuro-Computing Lecture 4 Radial Basis Function Network
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Shashi Shekhar Weili Wu Sanjay Chawla Ranga Raju Vatsavai
Support Vector Machines
Introduction to Radial Basis Function Networks
Empirical risk minimization
Neuro-Computing Lecture 2 Single-Layer Perceptrons
1st Annual Israel Multinational BMD Conference & Exhibition
What is The Optimal Number of Features
University of Wisconsin - Madison
Discriminative Training
Presentation transcript:

An M-ary KMP Classifier for Multi-aspect Target Classification Xuejun Liao, Hui Li, and Balaji Krishnapuram Department of Electrical and Computer Engineering Duke University International Conference on Acoustics, Speech, and Signal Processing, May l 7-21, 2004, Montreal, Quebec, Canada

Outline The M-ary Kernel Matching Pursuit (KMP) Multi-aspect High Range Resolution (HRR) Target Classification Results on MSTAR Dataset

The M-ary Classification Problem data Class 1  Class 2 Class m Class M binary classifier labeled as 0 One-versus-all rule: labeled as 1

The M-ary Classification Problem (Cont.) Training data set yi is the class label of xi is the importance weight of One-versus-all (OVA) re-labeling OVA labels

M-ary Kernel Matching Pursuit (KMP) Decision functions , m=1,2,…,M basis functions weight vectors Squared error Objective Find the  and {w[m]} that minimize e((x),{w[m]})

Least-Square Solution of w[m] where Substitute w[m] back into

Selection of Basis Functions Globally optimal selection Locally optimal selection Based on existing basis functions select the next one where is the error decrease due to 

Selection of Basis Functions (Cont.) Simple expression of the error decrease where

Selection of Basis Functions (Cont.) Update of the M-ary KMP Time complexity of the M-ary KMP Nc = number of candidate basis functions which is independent of M

Multi-aspect High Range Resolution (HRR) Target Classification Sensor-target configuration Depression angle Ground target Airborne radar sensor Multi-aspect HRR data with pose information xJ at J ... S2 S1 S3 S4 S7 S5 S6 S8 x2 at 2 x1 at 1 azimuth angle 

Probabilistic partition of azimuth and data State 3 Azimuth partition State 2 State 1 State 4 S2 S1 S3 S4 S7 S5 S6 S8 State 8 State 5 State 7 State 6 Data partition Di is the data set of state i, i=1,2,…, L

Multi-aspect decision function where is the M-ary KMP trained with the data in state i

The MSTAR dataset The ten MSTAR targets The data set consists of the X-band HRR signatures of the ten MSTAR targets. The HRR signatures have a range resolution of approximately 0.3 meters. Training data and testing data are distinct, each having a full coverage of 360 azimuth angles (depression angle fixed at 5), with 0.1 azimuth sampling

Results of the M-ary KMP Confusion matrix of KMP, with L=120 states in azimuth, test sequence spanning 3 azimuth. Sparsity:

Comparison to relevance vector machine (RVM) Confusion matrix of RVM, with L=120 states in azimuth, test sequence spanning 3 azimuth. Sparsity:

Classification performance versus azimuth span of observations

Convergence of the M-ary KMP Shown for the M-ary KMP in a single state

Conclusions The KMP is extended to the M-ary classification, with a time complexity which is independent of M The probabilistic state partition in multi-aspect HRR classification is handled in a natural way by the importance weights . The results on ten MSTAR targets show that at a comparable classification rate, the M-ary KMP achieves greater sparsity and shorter training time than the RVM.

Acknowledgment This work has been carried out under the supervision of Prof. Lawrence Carin of Duke University