Review of Statistical Pattern Recognition

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Introduction to Support Vector Machines (SVM)
1 Image Classification MSc Image Processing Assignment March 2003.
Curse of Dimensionality Prof. Navneet Goyal Dept. Of Computer Science & Information Systems BITS - Pilani.
DIMENSIONALITY REDUCTION: FEATURE EXTRACTION & FEATURE SELECTION Principle Component Analysis.
Detecting Faces in Images: A Survey
An Overview of Machine Learning
A Comprehensive Study on Third Order Statistical Features for Image Splicing Detection Xudong Zhao, Shilin Wang, Shenghong Li and Jianhua Li Shanghai Jiao.
Chapter 1: Introduction to Pattern Recognition
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Pattern Recognition 9/23/2008 Instructor: Wen-Hung Liao, Ph.D.
Chapter 2: Pattern Recognition
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Measuring Model Complexity (Textbook, Sections ) CS 410/510 Thurs. April 27, 2007 Given two hypotheses (models) that correctly classify the training.
Principle of Locality for Statistical Shape Analysis Paul Yushkevich.
Multidimensional Analysis If you are comparing more than two conditions (for example 10 types of cancer) or if you are looking at a time series (cell cycle.
Implementing a reliable neuro-classifier
A Kernel-based Support Vector Machine by Peter Axelberg and Johan Löfhede.
Lecture #1COMP 527 Pattern Recognition1 Pattern Recognition Why? To provide machines with perception & cognition capabilities so that they could interact.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Introduction to machine learning
Classification III Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart Russell,
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
PCA & LDA for Face Recognition
Introduction to Pattern Recognition
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Cyber Intrusion Detection Algorithm Based on Bayes’ Theorem
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
This week: overview on pattern recognition (related to machine learning)
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
The Perceptron. Perceptron Pattern Classification One of the purposes that neural networks are used for is pattern classification. Once the neural network.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
Optimal Bayes Classification
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Data Mining and Decision Support
Support-Vector Networks C Cortes and V Vapnik (Tue) Computational Models of Intelligence Joon Shik Kim.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Chapter 10 The Support Vector Method For Estimating Indicator Functions Intelligent Information Processing Laboratory, Fudan University.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
Finding Clusters within a Class to Improve Classification Accuracy Literature Survey Yong Jae Lee 3/6/08.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
CS 9633 Machine Learning Support Vector Machines
Machine Learning for Computer Security
Nonparametric Density Estimation – k-nearest neighbor (kNN) 02/20/17
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
PCA vs ICA vs LDA.
K Nearest Neighbor Classification
A Hybrid PCA-LDA Model for Dimension Reduction Nan Zhao1, Washington Mio2 and Xiuwen Liu1 1Department of Computer Science, 2Department of Mathematics Florida.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CSSE463: Image Recognition Day 14
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Outline A. M. Martinez and A. C. Kak, “PCA versus LDA,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp , 2001.
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Linear Discrimination
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
Hairong Qi, Gonzalez Family Professor
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Review of Statistical Pattern Recognition Wen-Hung Liao 9/22/2009

Review Paper A.K. Jain, R.P.W. Duin and J. Mao, “Statistical Pattern Recognition: A Review”, IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), Vol. 22, No. 1, pp. 4-37, Jan. 2000. More review papers: http://www.ph.tn.tudelft.nl/PRInfo/revpapers.html

Statistical Approach in PR Each pattern is represented in terms of d features and is viewed as a point in a d-dimensional feature space. Goal: establish decision boundaries to separate patterns belonging to different classes. Need to specify/estimate the probability distributions of the patterns.

Various Approaches in Statistical PR

Links Between Statistical and Neural Network Methods Linear Discriminant Function Principal Component Analysis Nonlinear Discriminant Function Parzen Window Density-based Classifier Perceptron Auto-Associative Networks Multilayer Perceptron Radial Basis Function Network

Model for Statistical Pattern Recognition Classification Feature Measurement Classification Preprocessing Training Feature Extraction /Selection Learning Preprocessing

The Curse of Dimensionality The performance of a classifier depends on the relationship between sample sizes, number of features and classifier complexity. Number of training data points be an exponential function of the feature dimension space.

Class-Conditional Probability Length d feature vector: x = (x1,x2,…,xd) C Classes (or categories): w1,w2,…,wc Class-conditional probability: The probability of x happening given that it belongs to class wi: p(x|wi)

How Many Features are Enough? Question: More features, better classification? Answer: Yes, if the class-conditional densities are completely known. No, if we need to estimate the the class-conditional densities.

Dimensionality Reduction Keep the number of features as small as possible (but not too small) due to: measurement cost classification accuracy Always some trade-off

Feature Extraction/Selection Feature Extraction: extract features from the sensed data Feature Selection: select (hopefully) the best subset of the input feature set. Feature extraction usually precedes selection Application-domain dependent

Example: Chernoff Faces Three classes of face Feature set: Nose length, mouth curvature, eye size, face shape. 150 4-d patterns, 50 patterns per class.

Chernoff Faces