Introduction to Pattern Recognition

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Perceptron Learning Rule
ECE 8443 – Pattern Recognition Objectives: Course Introduction Typical Applications Resources: Syllabus Internet Books and Notes D.H.S: Chapter 1 Glossary.
Machine Learning Neural Networks
Chapter 1: Introduction to Pattern Recognition
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
OUTLINE Course description, What is pattern recognition, Cost of error, Decision boundaries, The desgin cycle.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Recognition 9/23/2008 Instructor: Wen-Hung Liao, Ph.D.
A Study of Approaches for Object Recognition
Pattern Classification, Chapter 1 1 Basic Probability.
Chapter 6: Multilayer Neural Networks
Artificial Neural Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Introduction to machine learning
Radial Basis Function Networks
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture Notes by Neşe Yalabık Spring 2011.
嵌入式視覺 Pattern Recognition for Embedded Vision Template matching Statistical / Structural Pattern Recognition Neural networks.
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Introduction to Pattern Recognition Charles Tappert Seidenberg School of CSIS, Pace University.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
: Chapter 1: Introduction 1 Montri Karnjanadecha ac.th/~montri Principles of Pattern Recognition.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
IBS-09-SL RM 501 – Ranjit Goswami 1 Basic Probability.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Teuvo Kohonen Dr. Eng., Emeritus Professor of the Academy of Finland; Academician Since the 1960s, Professor Kohonen has introduced several new concepts.
Perception Introduction Pattern Recognition Image Formation
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Chapter 9 Neural Network.
Computational Intelligence II Lecturer: Professor Pekka Toivanen Exercises: Nina Rogelj
Compiled By: Raj G Tiwari.  A pattern is an object, process or event that can be given a name.  A pattern class (or category) is a set of patterns sharing.
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Introduction to Pattern Recognition (การรู้จํารูปแบบเบื้องต้น)
Data Mining and Decision Support
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 2 Nanjing University of Science & Technology.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Introduction to Pattern Recognition Chapter 1 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis 1.
Big data classification using neural network
Artificial Intelligence
Machine Learning for Computer Security
Pattern Classification Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 Dr. Ding Yuxin Pattern Recognition.
Learning in Neural Networks
Introduction Machine Learning 14/02/2017.
Data Mining, Neural Network and Genetic Programming
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
LECTURE 01: COURSE OVERVIEW
CS668: Pattern Recognition Ch 1: Introduction
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
Introduction to Pattern Recognition and Machine Learning
An Introduction to Supervised Learning
Neuro-Computing Lecture 4 Radial Basis Function Network
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Pattern Recognition
Creating Data Representations
Artificial Intelligence Lecture No. 28
Department of Electrical Engineering
LECTURE 01: COURSE OVERVIEW
Clustering Wei Wang.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Introduction to Neural Network
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

Introduction to Pattern Recognition

Introduction to Pattern Recognition Machine Perception An Example Pattern Recognition Systems The Design Cycle Learning and Adaptation Conclusion

What is Pattern… “A pattern is the opposite of a chaos; it is an entity vaguely defined, that could be given a name.” (Watanabe)

…Recognition? Identification of a pattern as a member of a category we already know, or we are familiar with. Clustering

Pattern Recognition Given an input pattern, make a decision about the “category” or “class” of the pattern Pattern recognition is a very broad subject with many applications In this course we will study a variety of techniques to solve P.R. problems and discuss their relative strengths and weaknesses

Pattern Class A collection of “similar” (not necessarily identical) objects A class is defined by class samples Inter-class variability Intra-class variability

Pattern Class Model Different descriptions, which are typically mathematical in form for each class/population Given a pattern, choose the best-fitting model for it and then assign it to class associated with the model

Intra-class and Inter-class Variability The letter “T” in different typefaces Same face under different expression, pose….

Interclass Similarity Characters that look similar Identical twins

Learning and Adaptation Supervised learning A teacher provides a category label or cost for each pattern in the training set Category “A” Category “B” Classification

Unsupervised learning The system forms clusters or “natural groupings” of the input patterns

Applications of Pattern Recognition Speech recognition Natural resource identification Character recognition (page readers, zip code, license plate) Identification and counting of cells Manufacturing Fingerprint identification Online handwriting retrieval …..

Pattern Recognition System Challenges Representation Matching A pattern recognition system involves Training Testing

Difficulties of Representation “How do you instruct someone (or some computer) to recognize caricatures in a magazine?” A representation could consist of a vector of real-valued numbers, ordered list of attributes, parts and their relations..

Good Representation Should have some invariant properties (e.g., w.r.t. rotation, translation, scale…) Account for intra-class variations Ability to discriminate pattern classes of interest Robustness to noise/occlusion Lead to simple decision making (e.g., decision boundary) Affordable

Pattern Recognition System Domain-specific knowledge Acquisition, representation Data acquisition Scanner, Camera, Ultrasound Preprocessing Image enhancement/restoration, segmentation Representation Features: color, shape, texture Decision making Statistical pattern recognition syntactic/structural pattern recognition Artificial neural networks Post-processing/Context

Arabic Optical Text Recognition System (AOTRS) Raw Image Select next Possible Segment Feature Extraction of possible segment Classification Arabic Text Post Processing Preprocessing Character Models Recognized? Yes No Adjust Possible Segment Possible Segmentation Points (PSP) Segment the recognized segment

Arabic text line

Pattern Recognition System Performance Error rate (Prob. of misclassification) on independent test samples Speed Cost Robustness Reject option Return On Investment (ROI)

An Example “Sorting incoming Fish on a conveyor according to species using optical sensing” Sea bass Species Salmon

Problem Analysis Set up a camera and take some sample images to extract features Length Lightness Width Number and shape of fins Position of the mouth, etc… This is the set of all suggested features to explore for use in our classifier!

Example (Cont.) Preprocessing Use a segmentation operation to isolate fishes from one another and from the background Information from a single fish is sent to a feature extractor whose purpose is to reduce the data by measuring certain features The features are passed to a classifier

Example (Cont.)

Example (Cont.) Classification Select the length of the fish as a possible feature for discrimination

Length Feature

Example (Cont.) The length is a poor feature alone! Select the lightness as a possible feature.

Lightness Feature

Task of decision theory Example (Cont.) Threshold decision boundary and cost relationship Move our decision boundary toward smaller values of lightness in order to minimize the cost (reduce the number of sea bass that are classified salmon!) Task of decision theory

Example (Cont.) Adopt the lightness and add the width of the fish Fish xT = [x1, x2] Lightness Width

Two-dimensional Feature Space (Representation) Cost of misclassification

Example (Cont.) We might add other features that are not correlated with the ones we already have. A precaution should be taken not to reduce the performance by adding such “noisy features” Ideally, the best decision boundary should be the one which provides an optimal performance such as in the following figure:

Complex Decision Boundary Issue of Generalization

Issue of generalization! Example (Cont.) However, our satisfaction is premature because the central aim of designing a classifier is to correctly classify novel input Issue of generalization!

Boundary With Good Generalization Simplify the decision boundary

Pattern Recognition Systems Sensing Use of a transducer (camera or microphone) PR system depends of the bandwidth, the resolution sensitivity distortion of the transducer Segmentation and grouping Patterns should be well separated and should not overlap

Pattern Recognition Systems (Cont.) Feature extraction Discriminative features Invariant features with respect to translation, rotation and scale. Classification Use a feature vector provided by a feature extractor to assign the object to a category Post Processing Exploit context input dependent information other than from the target pattern itself to improve performance

Features and model

Classification

The Design Cycle Data collection Feature Choice Model Choice Training Evaluation Computational Complexity

Data Collection How do we know when we have collected an adequately large and representative set of examples for training and testing the system?

Feature Choice Depends on the characteristics of the problem domain. Simple to extract, invariant to irrelevant transformation, insensitive to noise. How many features and which ones to use in constructing the decision boundary? Some features may be redundant! Curse of dimensionality —problems with too many features especially when we have a small number of training samples

Model Choice Model Choice Training Evaluation Unsatisfied with the performance of our fish classifier and want to jump to another class of model Training Use data to determine the classifier. Many different procedures for training classifiers and choosing models Evaluation Measure the error rate (or performance and switch from one set of features to another one)

Computational Complexity What is the trade-off between computational ease and performance? How an algorithm scales as a function of the number of features, patterns or categories?

Learning and Adaptation Supervised learning A teacher provides a category label or cost for each pattern in the training set Category “A” Category “B” Classification

Unsupervised learning The system forms clusters or “natural groupings” of the input patterns

Fingerprint Classification Assign fingerprints into one of pre-specified types

Fingerprint Enhancement To address the problem of poor quality fingerprints Noisy image Enhanced image

Fruit Sorter cherries apples grapefruits lemons

General Purpose P.R. System Humans have the ability to switch rapidly and seamlessly between different pattern recognition tasks It is very Difficult to design a device that is capable of performing a variety of classification tasks

Models for Pattern Recognition Template matching Statistical Syntactic (structural) Artificial neural network (biologically motivated?) …… Hybrid approach

Template Matching Template Input scene

Prototype registration to the low-level segmented image Deformable Template Prototype registration to the low-level segmented image Shape training set Prototype and variation learning Prototype warping

Statistical Pattern Recognition Patterns represented in a feature space Statistical model for pattern generation in feature space Given training patterns from each class, goal is to partition the feature space.

Approaches to Statistical Pattern Recognition

Statistical Pattern Recognition Preprocessing Feature extraction Classification Recognition Training Preprocessing Feature selection Learning Patterns + Class labels

Representation Each pattern is represented as a point in the d-dimensional feature space Features are domain-specific and be invariant to translation, rotation and scale Good representation  small intraclass variation, large interclass separation, simple decision rule

Invariant Representation Invariance to Translation Rotation Scale Skew Deformation

Structural Patten Recognition Decision-making when features are non-numeric or structural Describe complicated objects in terms of simple primitives and structural relationship Y N M L T X Z Scene Object Background D E

Syntactic Pattern Recognition Preprocessing Primitive, relation extraction Syntax, structural analysis pattern Recognition Training Preprocessing Primitive selection Grammatical, structural inference Patterns + Class labels

Chromosome Grammars Terminals: VT={,,,, } Non-terminals: VN={A,B,C,D,E,F} Pattern Classes:

Chromosome Grammars Image of human chromosomes Hierarchical-structure description of a submedium chromosome

Artificial Neural Networks Massive parallelism is essential for complex pattern recognition tasks (e.g., speech and image recognition) Human take only a few hundred ms for most cognitive tasks; suggests parallel computation Biological networks attempt to achieve good performance via dense interconnection of simple computational elements (neurons) Number of neurons  1010 – 1012 Number of interconnections/neuron  103 – 104 Total number of interconnections  1014

Artificial Neural Networks Nodes in neural networks are nonlinear, typically analog where is internal threshold or offset x1 w1 x2 Y (output) xd wd

Multilayer Perceptron Feed-forward nets with one or more layers (hidden) between the input and output nodes A three-layer net can generate arbitrary complex decision regions These nets can be trained by back-propagation training algorithm . . . . c outputs First hidden layer NH1 input units Second hidden layer NH2 input units d inputs

How m ch info mation are y u mi sing Utilizing Context How m ch info mation are y u mi sing Qvest

Comparing Pattern Recognition Models Template Matching Assumes very small intra-class variability Learning is difficult for deformable templates Syntactic Primitive extraction is sensitive to noise Describing a pattern in terms of primitives is difficult Statistical Assumption of density model for each class Neural Network Parameter tuning and local minima in learning In practice, statistical and neural network approaches work well

Summary Pattern recognition is extremely useful for Automatic decision making Assisting human decision makers Pattern recognition is a very difficult problem Successful systems have been built in well-constrained domains No single technique/model is suited for all pattern recognition problems Use of object models, constraints, and context is necessary for identifying complex patterns Careful sensor design and feature extraction can lead to simple classifiers

Key Concepts Pattern class Representation Feature selection\extraction Invariance (rotation, translation, scale, deformation) Preprocessing Segmentation Training samples Test samples Error rate Reject rate Curse of dimensionality

Key Concepts (Cont.) Supervised classification Decision boundary Unsupervised classification (clustering) Density Estimation Cost of misclassification/Risk Feature space partitioning Generalization/over fitting Contextual information Multiple classifiers Prior knowledge

Conclusion Reader seems to be overwhelmed by the number, complexity and magnitude of the sub-problems of Pattern Recognition Many of these sub-problems can indeed be solved Many fascinating unsolved problems still remain