Machine Learning on Images Janez Brank. Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How.

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints David Lowe.
Advertisements

ECG Signal processing (2)
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
An Introduction of Support Vector Machine
Support Vector Machines
CS4670 / 5670: Computer Vision Bag-of-words models Noah Snavely Object
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Discriminative and generative methods for bags of features
Effective Image Database Search via Dimensionality Reduction Anders Bjorholm Dahl and Henrik Aanæs IEEE Computer Society Conference on Computer Vision.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Content-based Image Retrieval CE 264 Xiaoguang Feng March 14, 2002 Based on: J. Huang. Color-Spatial Image Indexing and Applications. Ph.D thesis, Cornell.
Chapter 5: Linear Discriminant Functions
Lecture 28: Bag-of-words models
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
Colorization by Example R. Irony, D. Cohen-Or, D. Lischinski Tel-Aviv University The Hebrew University of Jerusalem Eurgraphics Symposium on Rendering.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
Image Categorization by Learning and Reasoning with Regions Yixin Chen, University of New Orleans James Z. Wang, The Pennsylvania State University Published.
Supervised Distance Metric Learning Presented at CMU’s Computer Vision Misc-Read Reading Group May 9, 2007 by Tomasz Malisiewicz.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Discriminative and generative methods for bags of features
Indexing Techniques Mei-Chen Yeh.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Computer Vision James Hays, Brown
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
CSE 185 Introduction to Computer Vision Pattern Recognition.
This week: overview on pattern recognition (related to machine learning)
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 24 – Classifiers 1.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Beyond Sliding Windows: Object Localization by Efficient Subwindow Search The best paper prize at CVPR 2008.
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Handwritten digit recognition
Click to edit Master subtitle style 2/23/10 Time and Space Optimization of Document Content Classifiers Dawei Yin, Henry S. Baird, and Chang An Computer.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Jakob Verbeek December 11, 2009
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Lecture 5: Statistical Methods for Classification CAP 5415: Computer Vision Fall 2006.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
SVMs in a Nutshell.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Face Detection EE368 Final Project Group 14 Ping Hsin Lee
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Instance Based Learning
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Histogram—Representation of Color Feature in Image Processing Yang, Li
LINEAR AND NON-LINEAR CLASSIFICATION USING SVM and KERNELS
K Nearest Neighbor Classification
Computer Vision Lecture 16: Texture II
Neuro-Computing Lecture 4 Radial Basis Function Network
Instance Based Learning
COSC 4335: Other Classification Techniques
Emel Doğrusöz Esra Ataer Muhammet Baştan Tolga Can
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Nearest Neighbors CSC 576: Data Mining.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

Machine Learning on Images Janez Brank

Introduction  Collections of images (a.k.a. pictorial databases)  Image retrieval  Image classification –How do we represent images? –How do we classify them based on this representation?

Global image descriptors (1)  Describe entire image with one vector  Histograms –Split (or “quantize”) the color space into C disjoint regions. –For each color region, remember what percentage of the image is covered by pixels whose colors are from that region.  Color coherence vectors (Pass et al., 1996, 1999) –Distinguish pixels belonging to a larger patch of pixels of the same color from the “lonelier” ones.

Global image descriptors (2)  Autocorrelograns (Huang et al., 1997) –Given a randomly chosen pixel of color c and a randomly chosen pixel at distance d from it, what is the probability that this second pixel also has color c? –Store this for all c and for some d.  Banded autocorrelograms (Huang et al., 1998) –For each c, sum the corresponding acg entries over all d.

Segmentation  Divide the image into several “regions”. Each should be uniform in terms of color and/or texture. –Describe each region separately. –How do we define similarity between images based on the similarity between regions?  WALRUS (Natsev et al., 1998) –Slice the image into small windows or tiles (e.g. 4  4 pixels) –Describe a window using the principal coefficients of the wavelet transform (  a 12-dimensional vector) –Cluster the descriptions (and hence windows).

Similarity measures for segmented images  WALRUS: –Look for pairs of regions (one region from each image) whose descriptions are close enough. –Calculate the percentage of the images covered by these regions.  IRM (integrated region matching; Li et al., 2000) –Distance between images is expressed as a weighted sum of distances between regions –To assign weights, similarity between regions, as well as their size, is taken into account.

Machine learning techniques  The k nearest neighbor method: –Classify a new instance by finding the k training instances closest to it and choosing the class prevailing among them.  Support vector machines: –Instances assumed to be vectors; separate two classes by placing a hyperplane between them. –By pretending to map the vectors into a different space (  kernel functions), we can achieve non-linear (e.g. polynomial) discrimination surfaces.

The database  1172 images, chosen from a larger collection, classified manually into 14 disjoint classes

The experiments  Global image descriptors: –histograms, coherence vectors, autocorrelograms, banded autocorrelograms –Several color space quantizations: RGB (64, 216, 256 regions), HSV (256 regions) –Nearest neighbors (using L 1 - ali L 2 -norm as the distance measure) –Support vector machines (three different kernels)  Segmentation: –Various segmentation parameters –Two similarity measures: WALRUS in IRM –Nearest neighbors

Poskusi z globalnimi opisi slik

Poskusi s segmentacijo

Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough

Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method

Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic and radial kernels better than linear

Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic, radial kernels better than linear  Segmentation parameters are important  IRM is better than the WALRUS similarity measure  Segmentation is not better than global descriptors

Conclusions  Autocorrelograms better than histograms, banded acg as good as full ones  The quantization of the color space should not be too rough  Support vector machines produce better classifiers than the nearest-neighbor method  Cubic, radial kernels better than linear  Segmentation parameters are important  IRM is better than the WALRUS similarity measure  Segmentation is not better than global descriptors  Looking at more nearest neigbors yields worse results

Zaključki  Avtokorelogrami boljši od histogramov, zgoščeni AKG enako dobri kot navadni  Metoda podpornih vektorjev daje boljše klasifikatorje od metode najbližjih sosedov  Metoda k najbližjih sosedov daje najboljše rezultate pri k = 1  Kvantizacija barvnega prostora ne sme biti pregroba  IRM daje boljše rezultate od WALRUS ove mere podobnosti  Klasifikatorji, dobljeni na podlagi segmentacije, niso nič boljši od tistih na podlagi globalnih opisov  S parametri segmentacije se da na točnost precej vplivati

Comparison with related work  Huang et al. (1998) built a sort of decision tree over a set of images, and also compared it with NN –Banded autocorrelograms, SVD (for each node of the tree) –Eleven classes, 90 images per class –Classification accuracies: around 80 %

Possibilities of future work  Understanding the unexpected results –The poor performance of segmentation-based classifiers –Why is k-NN best when k = 1?  Compare with other techniques –Other segmentation methods –Other color spaces  Looking for new image representation methods and new similarity measures