Download presentation
Presentation is loading. Please wait.
Published byLiliana Williams Modified over 6 years ago
1
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp , November, 1998.
2
Invariant Object Recognition
The central goal of computer vision research is to detect and recognize objects invariant to scale, viewpoint, illumination, and other changes November 21, 2018 Computer Vision
3
(Invariant) Object Recognition
November 21, 2018 Computer Vision
4
Generalization Performance
Many classifiers are available Maximum likelihood estimation, Bayesian estimation, Parzen Windows, Kn-nearest neighbor, discriminant functions, support vector machines, neural networks, decision trees, Which method is the best to classify unseen test data? The performance is often determined by features In addition, we are interested in systems that can solve a particular problem well November 21, 2018 Computer Vision
5
Error Rate on Hand Written Digit Recognition
November 21, 2018 Computer Vision
6
No Free Lunch Theorem November 21, 2018 Computer Vision
7
No Free Lunch Theorem – cont.
November 21, 2018 Computer Vision
8
Ugly Duckling Theorem In the absence of prior information, there is no principled reason to prefer one representation over another. November 21, 2018 Computer Vision
9
Bias and Variance Dilemma
Regression Find an estimate of a true but unknown function F(x) based on n samples generated by F(x) Bias – the difference between the expected value and the true value; a low bias means on average we will accurately estimate F from D Variance – the variability of estimation; a low bias means that the estimate does not change much as the training set varies. November 21, 2018 Computer Vision
10
Bias-Variance Dilemma
When the training data is finite, there is an intrinsic problem of any classifier function If the function is very generic, i.e., a non-parametric family, it suffers from high variance If the function is very specific, i.e., a parametric family, it suffers from high bias The central problem is to design a family of classifiers a priori such that both the variance and bias are low November 21, 2018 Computer Vision
11
November 21, 2018 Computer Vision
13
Bias and Variance vs. Model Complexity
November 21, 2018 Computer Vision
14
Gap Between Training and Test Error
Typically the performance of a classifier on a disjoint test set will be larger than that on the training set Where P is the number of training examples, h a measure of capacity (model complexity), a between 0.5 and 1, and k a constant November 21, 2018 Computer Vision
15
Check Reading System November 21, 2018 Computer Vision
16
End-to-End Training November 21, 2018 Computer Vision
17
Graph Transformer Networks
November 21, 2018 Computer Vision
18
Training Using Gradient-Based Learning
A multiple module system can be trained using a gradient-based method Similar to backpropagation used for multiple layer perceptrons November 21, 2018 Computer Vision
20
Convolutional Networks
November 21, 2018 Computer Vision
21
Handwritten Digit Recognition Using a Convolutional Network
November 21, 2018 Computer Vision
22
Training a Convolutional Network
The loss function used is Training algorithm is stochastic diagonal Levenberg-Marquardt RBF output is given by November 21, 2018 Computer Vision
23
MNIST Dataset 60,000 training images 10,000 test images
There are several different versions of the dataset November 21, 2018 Computer Vision
24
Experimental Results November 21, 2018 Computer Vision
25
Experimental Results November 21, 2018 Computer Vision
26
Distorted Patterns By using distorted patterns, the training error dropped to 0.8% from 0.95% without deformation November 21, 2018 Computer Vision
27
Misclassified Examples
November 21, 2018 Computer Vision
28
Comparison November 21, 2018 Computer Vision
29
Rejection Performance
November 21, 2018 Computer Vision
30
Number of Operations Unit: Thousand operations November 21, 2018
Computer Vision
31
Memory Requirements November 21, 2018 Computer Vision
32
Robustness November 21, 2018 Computer Vision
33
Convolutional Network for Object Recognition
November 21, 2018 Computer Vision
34
NORB Dataset November 21, 2018 Computer Vision
35
Convolutional Network for Object Recognition
November 21, 2018 Computer Vision
36
Experimental Results November 21, 2018 Computer Vision
37
Jittered Cluttered Dataset
November 21, 2018 Computer Vision
38
Experimental Results November 21, 2018 Computer Vision
39
Face Detection November 21, 2018 Computer Vision
40
Face Detection November 21, 2018 Computer Vision
41
Multiple Object Recognition
Based on heuristic over segmentation It avoids making hard decisions about segmentation by taking a large number of different segmentations November 21, 2018 Computer Vision
42
Graph Transformer Network for Character Recognition
November 21, 2018 Computer Vision
43
Recognition Transformer and Interpretation Graph
November 21, 2018 Computer Vision
44
Viterbi Training November 21, 2018 Computer Vision
45
Discriminative Viterbi Training
46
Discriminative Forward Training
November 21, 2018 Computer Vision
47
Space Displacement Neural Networks
By considering all possible locations, one can avoid explicit segmentation Similar to detection and recognition November 21, 2018 Computer Vision
48
Space Displacement Neural Networks
We can replicate convolutional networks at all possible locations November 21, 2018 Computer Vision
49
Space Displacement Neural Networks
November 21, 2018 Computer Vision
50
Space Displacement Neural Networks
November 21, 2018 Computer Vision
51
Space Displacement Neural Networks
November 21, 2018 Computer Vision
52
SDNN/HMM System November 21, 2018 Computer Vision
53
Graph Transformer Networks and Transducers
November 21, 2018 Computer Vision
54
On-line Handwriting Recognition System
November 21, 2018 Computer Vision
55
On-line Handwriting Recognition System
November 21, 2018 Computer Vision
56
Comparative Results November 21, 2018 Computer Vision
57
Check Reading System November 21, 2018 Computer Vision
58
Confidence Estimation
November 21, 2018 Computer Vision
59
Summary By carefully designing systems with desired invariance properties, one can often achieve better generalization performance by limiting system’s capacity Multiple module systems can be trained often effectively using gradient-based learning methods Even though in theory local gradient-based methods are subject to local minima, in practice it seems it is not a serious problem Incorporating contextual information into recognition systems are often critical for real world applications End-to-end training is often more effective November 21, 2018 Computer Vision
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.