Presentation is loading. Please wait.

Presentation is loading. Please wait.

Machine Learning on Images Janez Brank. Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How.

Similar presentations


Presentation on theme: "Machine Learning on Images Janez Brank. Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How."— Presentation transcript:

1 Machine Learning on Images Janez Brank

2 Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How do we represent images? –How do we classify them based on this representation?

3 Global image descriptors (1)  Describe entire image with one vector  Histograms –Split (or “quantize”) the color space into C disjoint regions. –For each color region, remember what percentage of the image is covered by pixels whose colors are from that region.  Color coherence vectors (Pass et al., 1996, 1999) –Distinguish pixels belonging to a larger patch of pixels of the same color from the “lonelier” ones.

4 Global image descriptors (2)  Autocorrelograns (Huang et al., 1997) –Given a randomly chosen pixel of color c and a randomly chosen pixel at distance d from it, what is the probability that this second pixel also has color c? –Store this for all c and for some d.  Banded autocorrelograms (Huang et al., 1998) –For each c, sum the corresponding acg entries over all d.

5 Segmentation  Divide the image into several “regions”. Each should be uniform in terms of color and/or texture. –Describe each region separately. –How do we define similarity between images based on the similarity between regions?  WALRUS (Natsev et al., 1998) –Slice the image into small windows or tiles (e.g. 4  4 pixels) –Describe a window using the principal coefficients of the wavelet transform (  a 12-dimensional vector) –Cluster the descriptions (and hence windows).

6 Similarity measures for segmented images  WALRUS: –Look for pairs of regions (one region from each image) whose descriptions are close enough. –Calculate the percentage of the images covered by these regions.  IRM (integrated region matching; Li et al., 2000) –Distance between images is expressed as a weighted sum of distances between regions –To assign weights, similarity between regions, as well as their size, is taken into account.

7 Machine learning techniques  The k nearest neighbor method: –Classify a new instance by finding the k training instances closest to it and choosing the class prevailing among them.  Support vector machines: –Instances assumed to be vectors; separate two classes by placing a hyperplane between them. –By pretending to map the vectors into a different space (  kernel functions), we can achieve non-linear (e.g. polynomial) discrimination surfaces.

8 The database  1172 images, chosen from a larger collection, classified manually into 14 disjoint classes

9 The experiments  Global image descriptors: –histograms, coherence vectors, autocorrelograms, banded autocorrelograms –Several color space quantizations: RGB (64, 216, 256 regions), HSV (256 regions) –Nearest neighbors (using L 1 - ali L 2 -norm as the distance measure) –Support vector machines (three different kernels)  Segmentation: –Various segmentation parameters –Two similarity measures: WALRUS in IRM –Nearest neighbors

10 Poskusi z globalnimi opisi slik

11

12 Poskusi s segmentacijo

13 Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough

14 Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method

15 Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic and radial kernels better than linear

16 Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic, radial kernels better than linear  Segmentation parameters are important  IRM is better than the WALRUS similarity measure  Segmentation is not better than global descriptors

17 Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic, radial kernels better than linear  Segmentation parameters are important  IRM is better than the WALRUS similarity measure  Segmentation is not better than global descriptors  Looking at more nearest neigbors yields worse results

18 Zaključki  Avtokorelogrami boljši od histogramov, zgoščeni AKG enako dobri kot navadni  Metoda podpornih vektorjev daje boljše klasifikatorje od metode najbližjih sosedov  Metoda k najbližjih sosedov daje najboljše rezultate pri k = 1  Kvantizacija barvnega prostora ne sme biti pregroba  IRM daje boljše rezultate od WALRUS ove mere podobnosti  Klasifikatorji, dobljeni na podlagi segmentacije, niso nič boljši od tistih na podlagi globalnih opisov  S parametri segmentacije se da na točnost precej vplivati

19 Comparison with related work  Huang et al. (1998) built a sort of decision tree over a set of images, and also compared it with NN –Banded autocorrelograms, SVD (for each node of the tree) –Eleven classes, 90 images per class –Classification accuracies: around 80 %

20 Possibilities of future work  Understanding the unexpected results –The poor performance of segmentation-based classifiers –Why is k-NN best when k = 1?  Compare with other techniques –Other segmentation methods –Other color spaces  Looking for new image representation methods and new similarity measures


Download ppt "Machine Learning on Images Janez Brank. Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How."

Similar presentations


Ads by Google