Download presentation
Presentation is loading. Please wait.
1
Segmentation
2
Terminology Segmentation, grouping, perceptual organization: gathering features that belong together Fitting: associating a model with observed features Top-down segmentation: pixels belong together because they come from the same object Bottom-up segmentation: pixels belong together because they look similar
3
The goals of segmentation Separate image into coherent “objects” Berkeley segmentation database: http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbench/ http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbench/ image human segmentation
4
The goals of segmentation Separate image into coherent “objects” Top-down or bottom-up process? Supervised or unsupervised? Group together similar-looking pixels for efficiency of further processing X. Ren and J. Malik. Learning a classification model for segmentation. ICCV 2003.Learning a classification model for segmentation. “superpixels”
5
The goals of segmentation Separate image into coherent “objects” Top-down or bottom-up process? Supervised or unsupervised? Group together similar-looking pixels for efficiency of further processing Related to image compression Measure of success is often application-dependent
6
Segmentation: Outline Inspiration from psychology Segmentation as clustering K-means Mean shift Segmentation as partitioning Graph-based segmentation, normalized cuts Integrating top-down and bottom-up segmentation for recognition
7
The Gestalt school Grouping is key to visual perception The Muller-Lyer illusion
8
The Gestalt school Grouping is key to visual perception Elements in a collection can have properties that result from relationships “The whole is greater than the sum of its parts”
9
The Gestalt school Grouping is key to visual perception Elements in a collection can have properties that result from relationships “The whole is greater than the sum of its parts” subjective contours occlusion familiar configuration http://en.wikipedia.org/wiki/Gestalt_psychology
10
Figure-ground discrimination
11
The ultimate Gestalt?
12
Gestalt factors These factors make intuitive sense, but are very difficult to translate into algorithms
13
Gestalt factors They may be hard to put into algorithms, but understanding them can come in useful for interface design
14
Gestalt factors They may be hard to put into algorithms, but understanding them can come in useful for interface design
15
Segmentation as clustering Source: K. Grauman
16
Segmentation as clustering Source: K. Grauman
17
Different clustering strategies Agglomerative clustering Start with each point in a separate cluster At each iteration, merge two of the “closest” clusters Divisive clustering Start with all points grouped into a single cluster At each iteration, split the “largest” cluster K-means clustering Iterate: assign points to clusters, compute means K-medoids Same as k-means, only cluster center cannot be computed by averaging The “medoid” of each cluster is the most centrally located point in that cluster (i.e., point with lowest average distance to the other points)
18
Image Intensity-based clustersColor-based clusters K-Means clustering K-means clustering based on intensity or color is essentially vector quantization of the image attributes Clusters don’t have to be spatially coherent
19
K-Means clustering K-means clustering based on intensity or color is essentially vector quantization of the image attributes Clusters don’t have to be spatially coherent Clustering based on (r,g,b,x,y) values enforces more spatial coherence
20
K-Means pros and cons Pros Simple and fast Converges to a local minimum of the error function Cons Need to pick K Sensitive to initialization Sensitive to outliers Only finds “spherical” clusters
21
http://www.caip.rutgers.edu/~comanici/MSPAMI/msPamiResults.html Mean shift segmentation An advanced and versatile technique for clustering-based segmentation D. Comaniciu and P. Meer, Mean Shift: A Robust Approach toward Feature Space Analysis, PAMI 2002.Mean Shift: A Robust Approach toward Feature Space Analysis
22
The mean shift algorithm seeks a mode or local maximum of density of a given distribution Choose a search window (width and location) Compute the mean of the data in the search window Center the search window at the new mean location Repeat until convergence Mean shift algorithm
23
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
24
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
25
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
26
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
27
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
28
Region of interest Center of mass Mean Shift vector Mean shift Slide by Y. Ukrainitz & B. Sarel
29
Region of interest Center of mass Mean shift Slide by Y. Ukrainitz & B. Sarel
30
Cluster: all data points in the attraction basin of a mode Attraction basin: the region for which all trajectories lead to the same mode Mean shift clustering Slide by Y. Ukrainitz & B. Sarel
31
Find features (color, gradients, texture, etc) Initialize windows at individual pixel locations Perform mean shift for each window until convergence Merge windows that end up near the same “peak” or mode Mean shift clustering/segmentation
32
http://www.caip.rutgers.edu/~comanici/MSPAMI/msPamiResults.html Mean shift segmentation results
33
More results
35
Mean shift pros and cons Pros Does not assume spherical clusters Just a single parameter (window size) Finds variable number of modes Robust to outliers Cons Output depends on window size Computationally expensive Does not scale well with dimension of feature space
36
Graph-based segmentation Represent features and their relationships using a graph Cut the graph to get subgraphs with strong interior links and weaker exterior links
37
Images as graphs Node for every pixel Edge between every pair of pixels (or every pair of “sufficiently close” pixels) Each edge is weighted by the affinity or similarity of the two nodes w ij i j Source: S. Seitz
38
Segmentation by graph partitioning Break Graph into Segments Delete links that cross between segments Easiest to break links that have low affinity –similar pixels should be in the same segments –dissimilar pixels should be in different segments ABC Source: S. Seitz w ij i j
39
Measuring Affinity Intensity Color Distance
40
Scale affects affinity Small σ: group only nearby points Large σ: group far-away points
41
Graph cut Set of edges whose removal makes a graph disconnected Cost of a cut: sum of weights of cut edges A graph cut gives us a segmentation What is a “good” graph cut and how do we find one? A B Source: S. Seitz
42
Graph cut
44
Affinity matrixBlock detection * Slides from Dan Klein, Sep Kamvar, Chris Manning, Natural Language Group Stanford University Multi-way graph cut
45
Minimum cut We can do segmentation by finding the minimum cut in a graph Efficient algorithms exist for doing this Drawback: minimum cut tends to cut off very small, isolated components * Slide from Khurram Hassan-Shafique CAP5415 Computer Vision 2003
46
Minimum cut We can do segmentation by finding the minimum cut in a graph Efficient algorithms exist for doing this Drawback: minimum cut tends to cut off very small, isolated components Ideal Cut Cuts with lesser weight than the ideal cut * Slide from Khurram Hassan-Shafique CAP5415 Computer Vision 2003
47
Normalized cut A minimum cut penalizes large segments This can be fixed by normalizing the cut by component size The normalized cut cost is: The exact solution is NP-hard but an approximation can be computed by solving a generalized eigenvalue problem assoc(A, V) = sum of weights of all edges in V that touch A J. Shi and J. Malik. Normalized cuts and image segmentation. PAMI 2000Normalized cuts and image segmentation.
48
Example results
50
Using texture features for segmentation Texture descriptor is vector of filter bank outputs J. Malik, S. Belongie, T. Leung and J. Shi. "Contour and Texture Analysis for Image Segmentation". IJCV 43(1),7-27,2001."Contour and Texture Analysis for Image Segmentation"
51
Using texture features for segmentation Texture descriptor is vector of filter bank outputs Textons are found by clustering J. Malik, S. Belongie, T. Leung and J. Shi. "Contour and Texture Analysis for Image Segmentation". IJCV 43(1),7-27,2001."Contour and Texture Analysis for Image Segmentation"
52
Using texture features for segmentation Texture descriptor is vector of filter bank outputs Textons are found by clustering Affinities are given by similarities of texton histograms over windows given by the “local scale” of the texture J. Malik, S. Belongie, T. Leung and J. Shi. "Contour and Texture Analysis for Image Segmentation". IJCV 43(1),7-27,2001."Contour and Texture Analysis for Image Segmentation"
53
The importance of scale J. Malik, S. Belongie, T. Leung and J. Shi. "Contour and Texture Analysis for Image Segmentation". IJCV 43(1),7-27,2001."Contour and Texture Analysis for Image Segmentation"
54
Example results J. Malik, S. Belongie, T. Leung and J. Shi. "Contour and Texture Analysis for Image Segmentation". IJCV 43(1),7-27,2001."Contour and Texture Analysis for Image Segmentation"
55
Pros Generic framework, can be used with many different features and affinity formulations Cons High storage requirement and time complexity Bias towards partitioning into equal segments * Slide from Khurram Hassan-Shafique CAP5415 Computer Vision 2003 Normalized cuts: Pro and con
56
Integrating top-down and bottom-up segmentation Z.W. Tu, X.R. Chen, A.L. Yuille, and S.C. Zhu. Image parsing: unifying segmentation, detection and recognition. IJCV 63(2), 113-140, 2005.Image parsing: unifying segmentation, detection and recognition.
57
Image parsing Define generative models for text and faces Deformable spline-based templates for characters PCA model for faces
58
Top-down: propose a model for a given region Bottom-up: verify the consistency of the model with image features Image parsing Data-driven Markov Chain Monte Carlo
59
Example results
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.