ENN: Extended Nearest Neighbor Method for Pattern Recognition

Slides:



Advertisements
Similar presentations
Principles of Density Estimation
Advertisements

On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach Author: Steven L. Salzberg Presented by: Zheng Liu.
COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –
Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
Naïve-Bayes Classifiers Business Intelligence for Managers.
Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
K-NEAREST NEIGHBORS AND DECISION TREE Nonparametric Supervised Learning.
Salvatore giorgi Ece 8110 machine learning 5/12/2014
Maximum Likelihood And Expectation Maximization Lecture Notes for CMPUT 466/551 Nilanjan Ray.
An Overview of Machine Learning
Classification and Decision Boundaries
ECE 8527 Homework Final: Common Evaluations By Andrew Powell.
Pattern recognition Professor Aly A. Farag
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Machine Vision and Dig. Image Analysis 1 Prof. Heikki Kälviäinen C50A6100 Lectures 12: Object Recognition Professor Heikki Kälviäinen Machine Vision and.
Prénom Nom Document Analysis: Non Parametric Methods for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Chapter 4 (part 2): Non-Parametric Classification
A Study of the Relationship between SVM and Gabriel Graph ZHANG Wan and Irwin King, Multimedia Information Processing Laboratory, Department of Computer.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Introduction to machine learning
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
1 Comparison of Discrimination Methods for the Classification of Tumors Using Gene Expression Data Presented by: Tun-Hsiang Yang.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Ajay Kumar, Member, IEEE, and David Zhang, Senior Member, IEEE.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Lecture note for Stat 231: Pattern Recognition and Machine Learning 4. Maximum Likelihood Prof. A.L. Yuille Stat 231. Fall 2004.
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
1 Generative and Discriminative Models Jie Tang Department of Computer Science & Technology Tsinghua University 2012.
Frontiers in the Convergence of Bioscience and Information Technologies 2007 Seyed Koosha Golmohammadi, Lukasz Kurgan, Brendan Crowley, and Marek Reformat.
1 E. Fatemizadeh Statistical Pattern Recognition.
A Bootstrap Interval Estimator for Bayes’ Classification Error Chad M. Hawes a,b, Carey E. Priebe a a The Johns Hopkins University, Dept of Applied Mathematics.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Neural Text Categorizer for Exclusive Text Categorization Journal of Information Processing Systems, Vol.4, No.2, June 2008 Taeho Jo* 報告者 : 林昱志.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 6: Nearest and k-nearest Neighbor Classification.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
1 Unsupervised Learning and Clustering Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Cost Modeling of Spatial Query Operators Using Nonparametric Regression Songtao Jiang Department of Computer Science University of Vermont October 10,
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Presenter : Chien-Shing Chen Author: Gustavo.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Big data classification using neural network
Data Transformation: Normalization
k-Nearest neighbors and decision tree
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Ch8: Nonparametric Methods
Overview of Supervised Learning
Boosting Nearest-Neighbor Classifier for Character Recognition
Statistical Learning Dong Liu Dept. EEIS, USTC.
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Nearest-Neighbor Classifiers
CAMCOS Report Day December 9th, 2015 San Jose State University
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

ENN: Extended Nearest Neighbor Method for Pattern Recognition This lecture notes is based on the following paper: B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015 ENN: Extended Nearest Neighbor Method for Pattern Recognition Prof. Haibo He Electrical Engineering University of Rhode Island, Kingston, RI 02881 Computational Intelligence and Self-Adaptive Systems (CISA) Laboratory http://www.ele.uri.edu/faculty/he/ Email: he@ele.uri.edu

Extended Nearest Neighbor for Pattern Recognition Limitations of K-Nearest Neighbors (KNN) “Two-way communication”: Extended Nearest Neighbors (ENN) Experimental Analysis Conclusion

Pattern Recognition Parametric Classifier Class-wise density estimation, including naive Bayes, mixture Gaussian, etc. Non-Parametric Classifier Nearest Neighbors Neural Network Support Vector Machine Nonparametric nature Easy implementation Powerfulness Robustness Consistency

Limitations of traditional KNN Scale-Sensitive Problem: The class 1 samples dominate their near neighborhood with higher density (i.e., more concentrated distribution). The class 2 samples are distributed in regions with lower density (i.e., more spread out distribution). Those class 2 samples which are close to the region of class 1 may be easily misclassified.

ENN: A New Approach Define generalized class-wise statistic for each class: Si denotes the samples in class i, and NNr(x, S) denotes the r-th nearest neighbor of x in S. Ti measures the coherence of data from the same class. 0 ≤ Ti ≤ 1 with Ti = 1 when all the nearest neighbors of class i data are also from the same class i, and with Ti = 0 when all the nearest neighbors are from other classes.

Intra-class coherence: Given an unknown sample Z to be classified, we iteratively assign it to class 1 and class 2, respectively, to obtain two new generalized class-wise statistics Tij, where j=1,2. Then, the sample Z is classified according to: ENN Classification Rule: Maximum Gain of Intra-class Coherence. For N-class classification:

To avoid the recalculation of generalized class-wise statistics in testing stage, an Equivalent Version of ENN is proposed:

The equivalent version has the same result as the original one, but avoids the recalculation of Tij

How this simple rule works better than KNN The ENN method makes a prediction in a “two-way communication” style: it considers not only who are the nearest neighbors of the test sample, but also who consider the test sample as their nearest neighbors.

Experimental Results and Analysis

Sampling methods Synthetic Data Set: A 3-dimensional Gaussian data with 3 classes: Considering the following four models, their error rates are: Model 2   Class 1 Class 2 Class 3 KNN ENN k = 3 32 31.9 39.3 34.4 31.4 30.5 k = 5 31.2 29.7 40.5 33.7 28.6 26.7 k = 7 28.5 28.3 40.8 33.6 25 24.3 Model 3 33.2 31 27 26.8 38.8 30.3 27.3 24 23.2 40.2 33.5 25.1 20.8 40.6 33

Sampling methods Real-life Data Sets: MNIST Handwritten Digit Recognition Data Examples

Sampling methods Real-life Data Sets: 20 data sets from UCI Machine Learning Repository t-test shows that ENN can significantly improve the classification performance in 17 out of 20 datasets, in comparison with KNN.

ENN Summary: Three versions of ENN B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015

ENN.V1 Summary: Three versions of ENN B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015

ENN.V2 Summary: Three versions of ENN B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015

Online Resources Summary: Three versions of ENN Supplementary materials and Matlab source code implementation available at: http://www.ele.uri.edu/faculty/he/research/ENN/ENN.html B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015

Conclusion A new ENN classification methodology based on the maximum gain of intra-class coherence. “Two-way communication”: ENN considers not only who are the nearest neighbors of the test sample, but also who consider the test sample as their nearest neighbors. Important and useful for many other machine learning and data mining problems, such as density estimation, clustering, regression, among others. B. Tang and H. He, "ENN: Extended Nearest Neighbor Method for Pattern Recognition," IEEE Computational Intelligence Magazine, vol.10, no.3, pp.52 - 60, Aug. 2015