An Introduction to Pattern Recognition

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
New Segmentation Technique
Presentation in Aircraft Satellite Image Identification Using Bayesian Decision Theory And Moment Invariants Feature Extraction Dickson Gichaga Wambaa.
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
An Overview of Machine Learning
Supervised Learning Recap
An Introduction to Sparse Coding, Sparse Sensing, and Optimization Speaker: Wei-Lun Chao Date: Nov. 23, 2011 DISP Lab, Graduate Institute of Communication.
G. Valenzise *, L. Gerosa, M. Tagliasacchi *, F. Antonacci *, A. Sarti * IEEE Int. Conf. On Advanced Video and Signal-based Surveillance, 2007 * Dipartimento.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Chapter 1: Introduction to Pattern Recognition
Lecture 20 Object recognition I
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
OUTLINE Course description, What is pattern recognition, Cost of error, Decision boundaries, The desgin cycle.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition 9/23/2008 Instructor: Wen-Hung Liao, Ph.D.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Machine Vision and Dig. Image Analysis 1 Prof. Heikki Kälviäinen C50A6100 Lectures 12: Object Recognition Professor Heikki Kälviäinen Machine Vision and.
Optimal Adaptation for Statistical Classifiers Xiao Li.
Lecture #1COMP 527 Pattern Recognition1 Pattern Recognition Why? To provide machines with perception & cognition capabilities so that they could interact.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
CEN 592 PATTERN RECOGNITION Spring Term CEN 592 PATTERN RECOGNITION Spring Term DEPARTMENT of INFORMATION TECHNOLOGIES Assoc. Prof.
Introduction Mohammad Beigi Department of Biomedical Engineering Isfahan University
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Isolated-Word Speech Recognition Using Hidden Markov Models
Introduction to Pattern Recognition Charles Tappert Seidenberg School of CSIS, Pace University.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Speech Recognition Pattern Classification. 22 September 2015Veton Këpuska2 Pattern Classification  Introduction  Parametric classifiers  Semi-parametric.
General Information Course Id: COSC6342 Machine Learning Time: TU/TH 10a-11:30a Instructor: Christoph F. Eick Classroom:AH123
1 TEMPLATE MATCHING  The Goal: Given a set of reference patterns known as TEMPLATES, find to which one an unknown pattern matches best. That is, each.
7-Speech Recognition Speech Recognition Concepts
DIGITAL IMAGE PROCESSING Dr J. Shanbehzadeh M. Hosseinajad ( J.Shanbehzadeh M. Hosseinajad)
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
INDEPENDENT COMPONENT ANALYSIS OF TEXTURES based on the article R.Manduchi, J. Portilla, ICA of Textures, The Proc. of the 7 th IEEE Int. Conf. On Comp.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Hierarchical Dirichlet Process (HDP) A Dirichlet process (DP) is a discrete distribution that is composed of a weighted sum of impulse functions. Weights.
Image Classification 영상분류
Hidden Markov Models in Keystroke Dynamics Md Liakat Ali, John V. Monaco, and Charles C. Tappert Seidenberg School of CSIS, Pace University, White Plains,
Maximum Entropy (ME) Maximum Entropy Markov Model (MEMM) Conditional Random Field (CRF)
PSEUDO-RELEVANCE FEEDBACK FOR MULTIMEDIA RETRIEVAL Seo Seok Jun.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
NTU & MSRA Ming-Feng Tsai
Rapporteur: 徐嘉駿 (Chia-Chun Hsu) Advisor: 丁建均 博士 (Jian-Jiun Ding, Ph.D) Graduate Institute of Communication Engineering National Taiwan University Taiwan,Taipei.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
Machine Learning: A Brief Introduction Fu Chang Institute of Information Science Academia Sinica ext. 1819
Pattern Recognition NTUEE 高奕豪 2005/4/14. Outline Introduction Definition, Examples, Related Fields, System, and Design Approaches Bayesian, Hidden Markov.
A Hybrid Model of HMM and RBFN Model of Speech Recognition 길이만, 김수연, 김성호, 원윤정, 윤아림 한국과학기술원 응용수학전공.
1 Review and Summary We have covered a LOT of material, spending more time and more detail on 2D image segmentation and analysis, but hopefully giving.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Deep Learning Amin Sobhani.
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Statistical Models for Automatic Speech Recognition
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Statistical Models for Automatic Speech Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Handwritten Characters Recognition Based on an HMM Model
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Presentation transcript:

An Introduction to Pattern Recognition Speaker : Wei–lun Chao Advisor : Prof. Jian-jiun Ding DISP Lab Graduate Institute of Communication Engineering National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Abstract Not a new research field Wide range included Enhancement by some factors: Computer architecture Machine learning Computer vision New way of thinking Improving human’s life National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Outline – What’s included What is pattern recognition Basic structure Different techniques Performance Care Example of applications Related works National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Content 1. Introduction 2. Basic Structure 3. Classification method I 4. Classification method II 5. Classification method III 6. Feature Generation 7. Feature Selection 8. Outstanding Application 9. Relation between IT and D&E 10. Conclusion National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 1. Introduction Pattern recognition is a process that taking in raw data and making an action based on the category of the pattern. What does a pattern means? “A pattern is essentially an arrangement”, N. Wiener [1] “A pattern is the opposite of a chaos”, Watanabe To be simplified, the interesting part National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

What can we do after analysis? Classification (Supervised learning) Clustering (Unsupervised learning) Other applications Category “A” Category “B” Classification Clustering National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Why we need pattern recognition? Human beings can easily recognize things or objects based on past learning experiences! Then how about computers? National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 2. Basic Structure Two basic factors: Feature & Classifier Feature: Car  Boundary Classifier: Mechanisms and methods to define what the pattern is National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan System structure The feature should be well-chosen to describe the pattern!! Knowledge: experience, analysis, trial & error The classifier should contain the knowledge of each pattern category and also the criterion or metric to discriminate among patterns classes. Knowledge : direct defined or “training“ National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Figure of system structure National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Four basic recognition models Template matching Syntactic Statistical Neural Network National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Another category idea Quantitative description: Using length, measure of area, and texture No relation between each component Structure descriptions: Qualitative factors Strings and trees Order, permutation, or hierarchical relations between each component National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

3. Classification method I Look-up table Decision-theoretic methods Distance Correlation Bayesian Classifier Neural network Popular methods nowadays National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 3.1 Bayesian classifier Two pattern classes: x is a pattern vector choose w1 for a specific x if P(w1|x)>P(w2|x) could be written as P(w1)P(x|w1)>P(w2)P(x|w2) based on the criterion to achieve the minimum overall error National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Bayesian classifier Multiple pattern classes: Risk based: conditional risk Minimum overall error based: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Bayesian classifier Decision function: A classifier assigns x to class wi if di(x)>dj(x) for all i ≠ j where di(x) are called decision (discriminant) functions Decision Boundary: The decision boundary between wi and wj for i ≠ j is that di(x)=dj(x) National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Bayesian classifier The most important point: probability model The widely-used model: Gaussian distribution for x is one-dimensional: for x is multi-dimensional: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 3.2 Neural network Without using statistical information Try to imitate how human learn A structure is generated based on perceptrons (hyperplane) National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Neural networks Multi-layer neural network National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Neural network What we need to define? Set the criterion for finding the best classifier Set the desired output Set the adapting mechanism The learning step: 1. Initialization: Assigning an arbitrary set of weights 2. Iterative step: Backward propagated modification 3. Stopping mechanism: Convergence under a threshold National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Neural network Complexity of Decision Surface Layer 1: line Layer 2: line intersection Layer 3: region intersection National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Popular methods nowadays Boosting: combining multiple learners Gaussian mixture model (GMM): Support vector machine (SVM): National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

4. Classification method II Template matching: There exists some relation between components of a pattern vector Methods: Measures based on correlation Computational consideration and improvement Measures based on optimal path searching techniques Deformable template matching National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

4.1 Measures based on correlation Distance: Normalized correlation: where i, j means the overlap region under translation Challenge: rotation, scaling, translation (RST) National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

4.2 Computational consideration and improvement Cross-correlation via its Fourier transform Direct computation: via the search window Improvement: Two-dimensional logarithmic search Hierarchical search Sequential methods National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

4.3 Measures based on optimal path searching techniques Pattern vectors are of different lengths Basic structure: Two-dimensional grid Elements of sequences on axes Each grid means correspondence between respective elements of the two sequences A path: Associated overall cost D: means the distance between respective elements of two strings National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Measures based on optimal path searching techniques Fast algorithm: Bellman’s principle the optimal path Necessary settings: Local constraint: Allowable transitions Global constraints: Dynamic programming End point constraints Cost measure: or National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

4.4 Deformable template matching Deformation parameters: Prototype A mechanism to deform the prototype A criterion to define the best match: -deformation parameter -matching energy -deformation energy National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

5. Classification method III Context-dependent methods: the class to which a feature vector is assigned depends (a) on its own value (b) on the values of the other feature vectors (c) on the existing relation among the various classes we have to consider more about the mutual information, which resides within the feature vectors Extension of the Bayesian classifier: N observations X: , M classes: and possible sequence National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Markov chain model First-order and two assumptions are made to simplify the task: We can get the probability terms: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan The Viterbi Algorithm Computational complexity: Direct way: Fast algorithm: Optimal path Cost function of a transition: The overall cost: Take the logarithm: Bellman’s principle: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Hidden Markov models Indirect observations of training data: Since the labeling has to obey the model structure Two cases: One model for (1) each class or (2) just an event Recognition: Assume we already know all PDF and types of states All path method: Each HMM could be described as: Best path method: Viterbi algorithm National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Training of HMM The most beautiful part of HMM For all path method: Baum-Welch re-estimation For best path method: Viterbi re-estimation Probability term: Discrete observation: Look-up table Continuous observation: Mixture model National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 6. Feature Generation Inability to use the raw data: (1) the raw data is too big to deal with (2) the raw data can’t give the classifier the same sense what people feel about the image National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 6.1 Regional feature First-order statistical features: mean, variance, skewness, kurtosis Second-order statistical features—Co-occurrence matrices National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Regional feature Local linear transforms for texture extraction Geometric moments: Zernike moments Parametric models: AR model National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 6.2 Shape & Size Boundary: Segmentation algorithm -> binarization -> and boundary extraction Invertible transform: Fourier transform Fourier-Mellin transform National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 6.2 Shape & Size Chain Codes: Moment-based features: Geometric moments National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 6.3 Audio feature Timbre: MFCC Rhythm: beat Melody: pitch National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 7. Feature Selection The main problem is the curse of dimensionality Reasons to reduce the number of features: Computational complexity: Trade-off between effectiveness & complexity Generalization properties: Related to the ratio of # training patterns to # classifier parameters Performance evaluation stage Basic criterion: Maintain large between-class distance and small within-class variance National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

8. Outstanding Application Speech recognition Movement recognition Personal ID Image retrieval by object query Camera & video recorder Remote sensing Monitoring …… National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Outstanding Application Retrieval: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Evaluation method P-R curve: Precision: a/c Recall: a/b a: # true got b: # retrieval c: # ground truth

9. Relation between IT and D&E Transmission: Pattern recognition: National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Graph of my idea National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan 10. Conclusion Pattern recognition is nearly everywhere in our life, each case relevant to decision, detection, retrieval can be a research topic of pattern recognition. The mathematics of pattern recognition is widely-inclusive, the methods of game theory, random process, decision and detection, or even machine learning. Feature cases: New features Better classifier Theory National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Idea of feature Different features perform well on different application: Ex: Video segmentation, video copy detection, video retrieval all use features from images (frame), while the features they use are different. Create new features National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

National Taiwan University, Taipei, Taiwan Idea of training Basic setting: Decision criterion Adaptation mechanism Initial condition Challenge: Insufficient training data Over-fitting National Taiwan University, Taipei, Taiwan DISP Lab @ MD531

Reference [1] R. C. Gonzalez, “Object Recognition,” in Digital image processing, 3rd ed. Pearson, August 2008, pp. 861-909. [2] Shyh-Kang Jeng, “Pattern recognition - Course Website,” 2009. [online] Available: http://cc.ee.ntu.edu.tw/~skjeng/PatternRecognition2007.htm. [Accessed Sep. 30, 2009]. [3] D. A. Forsyth, “CS 543 Computer Vision," Jan. 2009. [Online]. Available: http://luthuli.cs.uiuc.edu/~daf/courses/CS5432009/index.html. [Accessed: Oct. 21, 2009]. [4] Ke-Jie Liao, “Image-based Pattern Recognition Principles,” August 2008. [online] Available: http://disp.ee.ntu.edu.tw/research.php. [Accessed Sep. 19, 2009]. [5] E. Alpaydin, Introduction to Machine Learning. The MIT Press, 2004. [6] S. Theodoridis, K. Koutroumbas, Pattern Recognition, 2nd ed. Academic Press, 2003. [7] A. Yuille, P. Hallinan, and D. Cohen, “Feature Extraction from Faces Using Deformable Templates,” Int’l J. Computer Vision, vol. 8, no. 2, pp. 99-111, 1992. [8] J.S. Boreczky, L.D. Wilcox, “A hidden Markov model framework for video segmentation using audio and image features," in Proc. Int. Conf. Acoustics, Speech, and Signal Processing (ICASSP-98), Vol. 6, Seattle, WA, May 1998. [9] Ming-Sui Lee, “Digital Image Processing - Course Website,” 2009. [online] Available: http://www.csie.ntu.edu.tw/~dip/. [Accessed Oct. 21, 2009]. [10] W. Hsu, “Multimedia Analysis and Indexing – Course Website,” 2009. [online] Available: http://www.csie.ntu.edu.tw/~winston/courses/mm.ana.idx/index.html. [Accessed Oct. 21, 2009]. [11] R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, ed. John Wiley & Sons, 2001.