Presentation is loading. Please wait.

Presentation is loading. Please wait.

Introduction to Pattern Recognition

Similar presentations


Presentation on theme: "Introduction to Pattern Recognition"— Presentation transcript:

1 Introduction to Pattern Recognition

2 Introduction to Pattern Recognition
Machine Perception An Example Pattern Recognition Systems The Design Cycle Learning and Adaptation Conclusion

3 What is Pattern… “A pattern is the opposite of a chaos; it is an entity vaguely defined, that could be given a name.” (Watanabe)

4 …Recognition? Identification of a pattern as a member of a category we already know, or we are familiar with. Clustering

5 Pattern Recognition Given an input pattern, make a decision about the “category” or “class” of the pattern Pattern recognition is a very broad subject with many applications In this course we will study a variety of techniques to solve P.R. problems and discuss their relative strengths and weaknesses

6 Pattern Class A collection of “similar” (not necessarily identical) objects A class is defined by class samples Inter-class variability Intra-class variability

7 Pattern Class Model Different descriptions, which are typically mathematical in form for each class/population Given a pattern, choose the best-fitting model for it and then assign it to class associated with the model

8 Intra-class and Inter-class Variability
The letter “T” in different typefaces Same face under different expression, pose….

9 Interclass Similarity
Characters that look similar Identical twins

10 Learning and Adaptation
Supervised learning A teacher provides a category label or cost for each pattern in the training set Category “A” Category “B” Classification

11 Unsupervised learning
The system forms clusters or “natural groupings” of the input patterns

12 Applications of Pattern Recognition
Speech recognition Natural resource identification Character recognition (page readers, zip code, license plate) Identification and counting of cells Manufacturing Fingerprint identification Online handwriting retrieval …..

13 Pattern Recognition System
Challenges Representation Matching A pattern recognition system involves Training Testing

14 Difficulties of Representation
“How do you instruct someone (or some computer) to recognize caricatures in a magazine?” A representation could consist of a vector of real-valued numbers, ordered list of attributes, parts and their relations..

15 Good Representation Should have some invariant properties (e.g., w.r.t. rotation, translation, scale…) Account for intra-class variations Ability to discriminate pattern classes of interest Robustness to noise/occlusion Lead to simple decision making (e.g., decision boundary) Affordable

16 Pattern Recognition System
Domain-specific knowledge Acquisition, representation Data acquisition Scanner, Camera, Ultrasound Preprocessing Image enhancement/restoration, segmentation Representation Features: color, shape, texture Decision making Statistical pattern recognition syntactic/structural pattern recognition Artificial neural networks Post-processing/Context

17 Arabic Optical Text Recognition System (AOTRS)
Raw Image Select next Possible Segment Feature Extraction of possible segment Classification Arabic Text Post Processing Preprocessing Character Models Recognized? Yes No Adjust Possible Segment Possible Segmentation Points (PSP) Segment the recognized segment

18 Arabic text line

19 Pattern Recognition System Performance
Error rate (Prob. of misclassification) on independent test samples Speed Cost Robustness Reject option Return On Investment (ROI)

20 An Example “Sorting incoming Fish on a conveyor according to species using optical sensing” Sea bass Species Salmon

21

22 Problem Analysis Set up a camera and take some sample images to extract features Length Lightness Width Number and shape of fins Position of the mouth, etc… This is the set of all suggested features to explore for use in our classifier!

23 Example (Cont.) Preprocessing
Use a segmentation operation to isolate fishes from one another and from the background Information from a single fish is sent to a feature extractor whose purpose is to reduce the data by measuring certain features The features are passed to a classifier

24 Example (Cont.)

25 Example (Cont.) Classification
Select the length of the fish as a possible feature for discrimination

26 Length Feature

27 Example (Cont.) The length is a poor feature alone!
Select the lightness as a possible feature.

28 Lightness Feature

29 Task of decision theory
Example (Cont.) Threshold decision boundary and cost relationship Move our decision boundary toward smaller values of lightness in order to minimize the cost (reduce the number of sea bass that are classified salmon!) Task of decision theory

30 Example (Cont.) Adopt the lightness and add the width of the fish
Fish xT = [x1, x2] Lightness Width

31 Two-dimensional Feature Space (Representation) Cost of misclassification

32 Example (Cont.) We might add other features that are not correlated with the ones we already have. A precaution should be taken not to reduce the performance by adding such “noisy features” Ideally, the best decision boundary should be the one which provides an optimal performance such as in the following figure:

33 Complex Decision Boundary Issue of Generalization

34 Issue of generalization!
Example (Cont.) However, our satisfaction is premature because the central aim of designing a classifier is to correctly classify novel input Issue of generalization!

35 Boundary With Good Generalization Simplify the decision boundary

36 Pattern Recognition Systems
Sensing Use of a transducer (camera or microphone) PR system depends of the bandwidth, the resolution sensitivity distortion of the transducer Segmentation and grouping Patterns should be well separated and should not overlap

37 Pattern Recognition Systems (Cont.)
Feature extraction Discriminative features Invariant features with respect to translation, rotation and scale. Classification Use a feature vector provided by a feature extractor to assign the object to a category Post Processing Exploit context input dependent information other than from the target pattern itself to improve performance

38 Features and model

39 Classification

40 The Design Cycle Data collection Feature Choice Model Choice Training
Evaluation Computational Complexity

41 Data Collection How do we know when we have collected an adequately large and representative set of examples for training and testing the system?

42 Feature Choice Depends on the characteristics of the problem domain. Simple to extract, invariant to irrelevant transformation, insensitive to noise. How many features and which ones to use in constructing the decision boundary? Some features may be redundant! Curse of dimensionality —problems with too many features especially when we have a small number of training samples

43 Model Choice Model Choice Training Evaluation
Unsatisfied with the performance of our fish classifier and want to jump to another class of model Training Use data to determine the classifier. Many different procedures for training classifiers and choosing models Evaluation Measure the error rate (or performance and switch from one set of features to another one)

44 Computational Complexity
What is the trade-off between computational ease and performance? How an algorithm scales as a function of the number of features, patterns or categories?

45 Learning and Adaptation
Supervised learning A teacher provides a category label or cost for each pattern in the training set Category “A” Category “B” Classification

46 Unsupervised learning
The system forms clusters or “natural groupings” of the input patterns

47 Fingerprint Classification
Assign fingerprints into one of pre-specified types

48 Fingerprint Enhancement
To address the problem of poor quality fingerprints Noisy image Enhanced image

49 Fruit Sorter cherries apples grapefruits lemons

50 General Purpose P.R. System
Humans have the ability to switch rapidly and seamlessly between different pattern recognition tasks It is very Difficult to design a device that is capable of performing a variety of classification tasks

51 Models for Pattern Recognition
Template matching Statistical Syntactic (structural) Artificial neural network (biologically motivated?) …… Hybrid approach

52 Template Matching Template Input scene

53 Prototype registration to the low-level segmented image
Deformable Template Prototype registration to the low-level segmented image Shape training set Prototype and variation learning Prototype warping

54 Statistical Pattern Recognition
Patterns represented in a feature space Statistical model for pattern generation in feature space Given training patterns from each class, goal is to partition the feature space.

55 Approaches to Statistical Pattern Recognition

56 Statistical Pattern Recognition
Preprocessing Feature extraction Classification Recognition Training Preprocessing Feature selection Learning Patterns + Class labels

57 Representation Each pattern is represented as a point in the
d-dimensional feature space Features are domain-specific and be invariant to translation, rotation and scale Good representation  small intraclass variation, large interclass separation, simple decision rule

58 Invariant Representation
Invariance to Translation Rotation Scale Skew Deformation

59 Structural Patten Recognition
Decision-making when features are non-numeric or structural Describe complicated objects in terms of simple primitives and structural relationship Y N M L T X Z Scene Object Background D E

60 Syntactic Pattern Recognition
Preprocessing Primitive, relation extraction Syntax, structural analysis pattern Recognition Training Preprocessing Primitive selection Grammatical, structural inference Patterns + Class labels

61 Chromosome Grammars Terminals: VT={,,,, }
Non-terminals: VN={A,B,C,D,E,F} Pattern Classes:

62 Chromosome Grammars Image of human chromosomes
Hierarchical-structure description of a submedium chromosome

63 Artificial Neural Networks
Massive parallelism is essential for complex pattern recognition tasks (e.g., speech and image recognition) Human take only a few hundred ms for most cognitive tasks; suggests parallel computation Biological networks attempt to achieve good performance via dense interconnection of simple computational elements (neurons) Number of neurons  1010 – 1012 Number of interconnections/neuron  103 – 104 Total number of interconnections  1014

64 Artificial Neural Networks
Nodes in neural networks are nonlinear, typically analog where is internal threshold or offset x1 w1 x2 Y (output) xd wd

65 Multilayer Perceptron
Feed-forward nets with one or more layers (hidden) between the input and output nodes A three-layer net can generate arbitrary complex decision regions These nets can be trained by back-propagation training algorithm . . . . c outputs First hidden layer NH1 input units Second hidden layer NH2 input units d inputs

66 How m ch info mation are y u mi sing
Utilizing Context How m ch info mation are y u mi sing Qvest

67 Comparing Pattern Recognition Models
Template Matching Assumes very small intra-class variability Learning is difficult for deformable templates Syntactic Primitive extraction is sensitive to noise Describing a pattern in terms of primitives is difficult Statistical Assumption of density model for each class Neural Network Parameter tuning and local minima in learning In practice, statistical and neural network approaches work well

68 Summary Pattern recognition is extremely useful for
Automatic decision making Assisting human decision makers Pattern recognition is a very difficult problem Successful systems have been built in well-constrained domains No single technique/model is suited for all pattern recognition problems Use of object models, constraints, and context is necessary for identifying complex patterns Careful sensor design and feature extraction can lead to simple classifiers

69 Key Concepts Pattern class Representation Feature selection\extraction
Invariance (rotation, translation, scale, deformation) Preprocessing Segmentation Training samples Test samples Error rate Reject rate Curse of dimensionality

70 Key Concepts (Cont.) Supervised classification Decision boundary
Unsupervised classification (clustering) Density Estimation Cost of misclassification/Risk Feature space partitioning Generalization/over fitting Contextual information Multiple classifiers Prior knowledge

71 Conclusion Reader seems to be overwhelmed by the number, complexity and magnitude of the sub-problems of Pattern Recognition Many of these sub-problems can indeed be solved Many fascinating unsolved problems still remain


Download ppt "Introduction to Pattern Recognition"

Similar presentations


Ads by Google