Download presentation
Presentation is loading. Please wait.
1
Principle of Locality for Statistical Shape Analysis Paul Yushkevich
2
Outline Classification Shape variability with components Principle of Locality Feature Selection for Classification Inter-scale residuals
3
Classification Given training data belonging to 2 or more classes, construct a classifier. A classifier assigns a class label to each point in space. Classifiers can be: –Parametric vs. Nonparametric –ML vs. MAP –Linear vs. Non-linear
4
Common Classifiers Fisher Linear Discriminant –Linear, Parametric (Gaussian model with pooled covariance) K Nearest Neighbor, Parzen Windows –Non-linear, non-parametric Support Vector Machines –Non-parametric, linear
5
Examples Fisher ClassifierNon-linear MLE Classifier 3-Nearest Neighbor
6
Shape Variability Multiple (Independent) Components –Local and Global –Coarse and Fine
7
Principle of Locality Heuristic about biological shape: –Each component affects a region of the object –Each component affects a range of scales
8
Feature selection If we limit classification to just the right subset of features, we can drastically improve the results
9
Feature Selection Example Underlying distributions Training Set sampled from the distributions Classifier constructed in 2 dimensions Classifier constructed in the relevant y-dimension
10
Feature Selection Example p=2, =1p=4, =1 p=4, =2p=8, =2 Expected error rate in classification in p dimensions vs. classification in the relevant dimensions, plotted against training sample size.
11
Synthetic Shapes
12
Feature Selection in Shape Relevant features should form a small set ‘windows’ The order of features is important Example: –Fisher classification rate over all possible windows Window position Window size
13
Feature Search Algorithm Look for best set of feature windows and best classifier simultaneously Similar to SVMs Minimize: E separation + E Nfeature + E Nwindows
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.