Presentation is loading. Please wait.

Presentation is loading. Please wait.

Image Segmentation A Graph Theoretic Approach. Factors for Visual Grouping Similarity (gray level difference) Similarity (gray level difference) Proximity.

Similar presentations


Presentation on theme: "Image Segmentation A Graph Theoretic Approach. Factors for Visual Grouping Similarity (gray level difference) Similarity (gray level difference) Proximity."— Presentation transcript:

1 Image Segmentation A Graph Theoretic Approach

2 Factors for Visual Grouping Similarity (gray level difference) Similarity (gray level difference) Proximity Proximity Continuity Continuity Reference: Reference: M. Wertheimer, “Laws of Organization in Perceptual Forms”, A Sourcebook of Gestalt Psychology, W.B. Ellis, ed., pp. 71-88, Harcourt, Brace, 1938.

3 What is the “correct” grouping?

4 Subjectivity in Segmentation Prior world knowledge needed Prior world knowledge needed Agglomerative and divisive techniques in grouping (or Region-based merge and split algorithms in image segmentation) Agglomerative and divisive techniques in grouping (or Region-based merge and split algorithms in image segmentation) Local properties – easier to specify but poorer results Local properties – easier to specify but poorer results e.g. coherence of brightness, colour, texture, motion Global properties – more difficult to specify but give better results e.g. object symmetries Global properties – more difficult to specify but give better results e.g. object symmetries Image segmentation can be modeled as a graph partitioning and optimization problem Image segmentation can be modeled as a graph partitioning and optimization problem

5 Partitioning Divisive or top-down approach Divisive or top-down approach Inherently hierarchical Inherently hierarchical We must aim at returning a tree structure (called the dendogram) corresponding to a hierarchical partitioning scheme instead of a single “flat” partition We must aim at returning a tree structure (called the dendogram) corresponding to a hierarchical partitioning scheme instead of a single “flat” partition

6 Challenges Picking an appropriate criterion to minimize which would result in a “good” segmentation Picking an appropriate criterion to minimize which would result in a “good” segmentation Finding an efficient way to achieve the minimization Finding an efficient way to achieve the minimization

7 Modeling as a Graph Partitioning problem Set of points of the feature space represented as a weighted, undirected graph, G = (V, E) Set of points of the feature space represented as a weighted, undirected graph, G = (V, E) The points of the feature space are the nodes of the graph. The points of the feature space are the nodes of the graph. Edge between every pair of nodes. Edge between every pair of nodes. Weight on each edge, w(i, j), is a function of the similarity between the nodes i and j. Weight on each edge, w(i, j), is a function of the similarity between the nodes i and j. Partition the set of vertices into disjoint sets where similarity within the sets is high and across the sets is low. Partition the set of vertices into disjoint sets where similarity within the sets is high and across the sets is low.

8 Weight Function for Brightness Images Weight measure (reflects likelihood of two pixels belonging to the same object) Weight measure (reflects likelihood of two pixels belonging to the same object)

9 Representing Images as Graphs

10 Graph Weight Matrix, W

11 Segmentation and Graphs - Other Common Approaches Minimal Spanning Tree Minimal Spanning Tree Limited Neighbourhood Set Limited Neighbourhood Set - Both approaches are computationally efficient but the criteria are based on local properties - Perceptual grouping is about extracting global impressions of a scene; thus local criteria are often inadequate

12 First attempt at global criterion selection A graph can be partitioned into two disjoint sets simply removing the edges connecting the two parts A graph can be partitioned into two disjoint sets by simply removing the edges connecting the two parts The degree of dissimilarity between these two pieces can be computed as total weight of the edges that have been removed More formally, it is called the ‘cut’

13 Graph Cut

14 Optimization Problem Minimize the cut value Minimize the cut value No of such partitions is exponential (2^N) but the minimum cut can be found efficiently No of such partitions is exponential (2^N) but the minimum cut can be found efficiently Reference: Z. Wu and R. Leahy, “An Optimal Graph Theoretic Approach to Data Clustering: Theory and Its Application to Image Segmentation”. IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 15, no. 11, pp. 1101-1113, Nov. 1993. Subject to the constraints:

15 Problems with min-cut Minimum cut criteria favors cutting small sets of isolated nodes in the graph.

16 Solution – Normalized Cut We must avoid unnatural bias for partitioning out small sets of points Normalized Cut - computes the cut cost as a fraction of the total edge connections to all the nodes in the graph where

17 Looking at it another way.. Our criteria can also aim to tighten similarity within the groups Our criteria can also aim to tighten similarity within the groups Minimizing Ncut and maximizing Nassoc are actually equivalent Minimizing Ncut and maximizing Nassoc are actually equivalent

18 Matrix Formulations Let x be an indicator vector s.t. x i = 1, if i belongs to A 0,otherwise 0,otherwise Assoc(A, A) = x T Wx Assoc(A, A) = x T Wx Assoc(A, V) = x T Dx Assoc(A, V) = x T Dx Cut(A, V-A) = x T (D – W)x Cut(A, V-A) = x T (D – W)x

19 Computational Issues Exact solution to minimizing normalized cut is an NP-complete problem Exact solution to minimizing normalized cut is an NP-complete problem However, approximate discrete solutions can be found efficiently However, approximate discrete solutions can be found efficiently Normalized cut criterion can be computed efficiently by solving a generalized eigenvalue problem Normalized cut criterion can be computed efficiently by solving a generalized eigenvalue problem

20 Algorithm 1. Construct the weighted graph representing the image. Summarize the information into matrices, W & D. Edge weight is an exponential function of feature similarity as well as distance measure. 2. Solve for the eigenvectors with the smallest eigenvalues of: (D – W)x = LDx

21 Algorithm (contd.) 3. Partition the graph into two pieces using the second smallest eigenvector. Signs tell us exactly how to partition the graph. 4. Recursively run the algorithm on the two partitioned parts. Recursion stops once the Ncut value exceeds a certain limit. This maximum allowed Ncut value controls the number of groups segmented.

22 Computational Issues Revisited Solving a standard eigenvalue problem for all eigenvectors takes O(n^3) operations, where n is the number of nodes in the graph This becomes impractical for image segmentation applications where n is the number of pixels in an image For the problem at hand, the graphs are often only locally connected, only the top few eigenvectors are needed for graph partitioning, and the precision requirement for the eigenvectors is low, often only the right sign bit is required.

23 A Physical Interpretation Think of the weighted graph as a spring mass system Think of the weighted graph as a spring mass system  Graph nodes  physical masses  Graph edges  springs  Graph edge weight  spring stiffness  Total incoming edge weights  mass of the node

24 A Physical Interpretation (contd..) Imagine giving a hard shake to this spring-mass system, forcing the nodes to oscillate in the direction perpendicular to the image plane Nodes that have stronger spring connections among them will likely oscillate together Eventually, the group will “pop” off from the image plane The overall steady state behavior of the nodes can be described by its fundamental mode of oscillation and it can be shown that the fundamental modes of oscillation of this spring mass system are exactly the generalized eigenvectors of the normalized cut.

25 Comparisons with other criteria Average Cut: Average Cut: Analogously, Average Association can be defined as: Analogously, Average Association can be defined as: Unlike in the case of Normalized Cut and Normalized Association, Average Cut and Average Association do not have a simple relationship between them Unlike in the case of Normalized Cut and Normalized Association, Average Cut and Average Association do not have a simple relationship between them Consequently, one cannot simultaneously minimize the disassociation across the partitions while maximizing the association within the groups Normalized Cut produces better results in practice Normalized Cut produces better results in practice

26 Comparisons with other criteria (contd..)

27 Average association has a bias for finding tight clusters – runs the risk of finding small, tight clusters in the data Average association has a bias for finding tight clusters – runs the risk of finding small, tight clusters in the data Average cut does not look at within-group similarity – problems when the dissimilarity between groups is not clearly defined Average cut does not look at within-group similarity – problems when the dissimilarity between groups is not clearly defined

28 Consider random 1-D data points: Consider random 1-D data points: Each data point is a node in the graph Each data point is a node in the graph and the weighted graph edge connecting two points is defined to be inversely proportional to the distance between two nodes We will consider two different monotonically decreasing weight functions, w(i,j) = f(d(i,j)), defined on the distance function, d(i,j), with differents rate of fall-off.

29 Fast falling weight function With this function, only close-by points are connected. With this function, only close-by points are connected.

30 Criterion usedSecond smallest eigenvector plot

31 Interpretation The cluster on the right has less within- group similarity compared with the cluster on the left. In this case, average association fails to find the right partition. Instead, it focuses on finding small clusters in each of the two main subgroups.

32 Slowly decreasing weight function With this function, most points have non- trivial connections with the rest With this function, most points have non- trivial connections with the rest

33 Criterion usedSecond smallest eigenvector plot

34 Interpretation To find a cut of the graph, a number of edges with heavy weights have to be removed. In this case, average cut has trouble deciding on where to cut.

35 Reference J. Shi and J. Malik, “Normalized Cuts and Image Segmentation,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 22, no. 8, pp. 888-905, Aug. 2000.


Download ppt "Image Segmentation A Graph Theoretic Approach. Factors for Visual Grouping Similarity (gray level difference) Similarity (gray level difference) Proximity."

Similar presentations


Ads by Google