Intrinsic Data Geometry from a Training Set

Slides:



Advertisements
Similar presentations
Partitional Algorithms to Detect Complex Clusters
Advertisements

1 Manifold Alignment for Multitemporal Hyperspectral Image Classification H. Lexie Yang 1, Melba M. Crawford 2 School of Civil Engineering, Purdue University.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Differential geometry I
Spectral methods 1 © Alexander & Michael Bronstein,
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
A Geometric Perspective on Machine Learning 何晓飞 浙江大学计算机学院 1.
Nonlinear Dimension Reduction Presenter: Xingwei Yang The powerpoint is organized from: 1.Ronald R. Coifman et al. (Yale University) 2. Jieping Ye, (Arizona.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Semi-supervised Learning Rong Jin. Semi-supervised learning  Label propagation  Transductive learning  Co-training  Active learning.
Presented by: Mingyuan Zhou Duke University, ECE April 3, 2009
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Graph Based Semi- Supervised Learning Fei Wang Department of Statistical Science Cornell University.
Principal Component Analysis
Spectral embedding Lecture 6 1 © Alexander & Michael Bronstein
Semi-Supervised Classification by Low Density Separation Olivier Chapelle, Alexander Zien Student: Ran Chang.
Normalized Cuts and Image Segmentation Jianbo Shi and Jitendra Malik, Presented by: Alireza Tavakkoli.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Spectral Embedding Alexander Bronstein, Michael Bronstein
A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford Presented by Napat Triroj.
Semi-Supervised Learning D. Zhou, O Bousquet, T. Navin Lan, J. Weston, B. Schokopf J. Weston, B. Schokopf Presents: Tal Babaioff.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Diffusion Maps and Spectral Clustering
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Dimensionality reduction Usman Roshan CS 675. Supervised dim reduction: Linear discriminant analysis Fisher linear discriminant: –Maximize ratio of difference.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
Introduction --Classification Shape ContourRegion Structural Syntactic Graph Tree Model-driven Data-driven Perimeter Compactness Eccentricity.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Lecture 19 Representation and description II
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
Recognition and Matching based on local invariant features Cordelia Schmid INRIA, Grenoble David Lowe Univ. of British Columbia.
Segmentation using eigenvectors Papers: “Normalized Cuts and Image Segmentation”. Jianbo Shi and Jitendra Malik, IEEE, 2000 “Segmentation using eigenvectors:
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
Chapter 14: SEGMENTATION BY CLUSTERING 1. 2 Outline Introduction Human Vision & Gestalt Properties Applications – Background Subtraction – Shot Boundary.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
GRASP Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger, Fei Sha and Lawrence K. Saul ICML’04 Department of Computer.
Manifold learning: MDS and Isomap
Nonlinear Dimensionality Reduction Approach (ISOMAP)
H. Lexie Yang1, Dr. Melba M. Crawford2
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
A Convergent Solution to Tensor Subspace Learning.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Introduction to Machine Learning Nir Ailon Lecture 12: EM, Clustering and More.
Spectral Methods for Dimensionality
Nonlinear Dimensionality Reduction
Random remarks about random walks
LECTURE 10: DISCRIMINANT ANALYSIS
Solving Linear Systems Ax=b
Unsupervised Riemannian Clustering of Probability Density Functions
Dimensionality reduction
Generating multidimensional embeddings based on fuzzy memberships
Spectral Methods Tutorial 6 1 © Maks Ovsjanikov
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Learning with information of features
REMOTE SENSING Multispectral Image Classification
REMOTE SENSING Multispectral Image Classification
Grouping.
EE513 Audio Signals and Systems
Spectral Clustering Eric Xing Lecture 8, August 13, 2010
LECTURE 09: DISCRIMINANT ANALYSIS
Using Manifold Structure for Partially Labeled Classification
Recognition and Matching based on local invariant features
Marios Mattheakis and Pavlos Protopapas
Presentation transcript:

Intrinsic Data Geometry from a Training Set Phd Thesis Proposal Nicola Rebagliati Supervisor: Alessandro Verri December 2007

Index Introduction The Training Set Intrinsic Geometry An interesting operator over manifolds The Laplace-Beltrami operator The Laplacian matrix Preliminary work Image segmentation Image signatures Work plan

Introduction

The Training Set In the context of Statistical Learning we are given a set of training data representative of the collection of all possible data: Where, in the classification case, That is: Vectors in a Euclidean space and Labels

The Training Set The labelling function is unknown We are looking for a function f* that minimize an error over future data We should stay away from overfitting and underfitting A solution... ... that overfits

Some example of Training Set -1 1 Images: R^10000 Faces Not Faces Patient information: R^15 Exceeding quantity of iron Right quantity of iron Actions: R^(10000 x k) Walking Resting

Intrinsic Dimensionality Without assumptions vectors may be uniformly distributed in their euclidean space In many practical cases the degrees of freedom of training data are far less than n. Images of 100X100 pixels representing squares may have three degrees of freedom, for example.

Data form a structure We may have something more than a simple lower dimensionality Data may live in a topological space where we can compute angles and distances That is a Riemannian Manifold

Example: The Swiss Roll Difference between euclidean distance and geodesic distance

Semisupervised Learning The geometry of the data is exploited in the semisupervised learning [BelkinNiyogi2004] In Semisupervised Learning the Training set is made of Labeled Data and Unlabeled Data Structural penalty w.r.t. known data Structural penalty w.r.t. all data Empirical error

Geometrical Information A manifold has some intrinsic geometrical information [McKeanSinger67] : The Volume The Volume of the Boundary The Euler Characteristic ... The training set is extracted from the manifold Is it possible to approximate these information with the training set?

An interesting operator over manifold

The continuos Case: The Laplace-Beltrami operator Let M denote a manifold The Laplace Beltrami operator maps functions to functions, both defined on the manifold In a while we see why we are interested in its eigenvalues and eigenvectors

The heat trace The heat trace of the Laplacian on a Riemannian manifold can be expressed as: The heat trace has an asymptotic expansion: C0 is the volume of the manifold C1 is proportional to the volume of the Boundary ... [McKeanSinger]

The discrete Case: The Laplacian Matrix The Training Set can be seen as an undirected weighted graph Choose a comparing function for weighting graph edges: Define the degree of a vertex as Consider the Weight matrix Consider the diagonal Degree matrix

The discrete Case: The Laplacian Matrix The Laplacian Matrix is simply: A normalized version [VonLuxburg06] has some better properties:

The discrete Case: The Laplacian Matrix Now consider the spectral decomposition of the normalized laplacian: The eigenvalues are ordered as The training set can be embedded in a possibly low dimensional space through the eigenvectors and eigenvalues:

The Convergence We would like the discrete operator to be consistent with the continuos one Pointwise distance between the two operator is bounded by an error: The error is divided in Bias + Variance Here we have some constants that depend on the possibly unknown underlying manifold

Image segmentation PRELIMINARY WORK

Image segmentation A pixel is given by a digital camera as a tuple (x, y, R, G, B) We may define features that map a pixel, usually considering also its neighborhood, on a vector in For example (x, y, R, G, B) -> (R,G,B)

Image segmentation Clustering methods group together different vectors I want to compare clustering results with a ‘ground truth’, or ‘human truth’

Segmenting with Laplacian It can be proved that the second eigenvector of the Laplacian matrix give a clustering that: Maximize intra-group cohesion Minimize inter-group cohesion I consider as weighting function the gaussian: What sigma should be used?

Choose sigma in image segmentation Heuristic: Choosing sigma such that the underlying manifold, if exists, is best approximated That is The training set is sampled over the manifold, so it is a random variable Choose sigma such that we have the minimum Bias-Variance error Let’s see a syntetic case

Choose sigma in image segmentation 3D Sphere Error Bound

Choose sigma and the balancing principle The error bound is divided in bias and variance Algorithm for balancing the two terms We do not know the constants

Image Signatures PRELIMINARY WORK

3D-shape retrieval in database In an article Reuter et al. [ReuterWolterPeinecke] propose to use the spectrum of the laplace-beltrami operator of a shape as its signature. This signature has many desiderable properties within that context: Representation invariant Scaling invariant Continuos w.r.t. Shape deformations

Image Signature In an another article Peinecke et al. [PeineckeReuterWolter] have proposed to look at an image as a 2D-manifold and to use its laplacian spectrum as signature. It is not clear what are the useful geometric invariants in that case: Illumination? Scale?

Image Signature: Laplacian Definition Different Kind of Laplacian Definition The Normalized Laplacian The Laplace-Kirchoff operator The continuos operator Computed using the Finite Element Method (FEM) Different Results

Laplace-Kirchoff operator, radius:1

Laplace operator, sigma_xy=30, sigma gray=50

Work to do

Short Term Investigate an empirical criterion or an algorithm for choosing a suitable sigma Tackle the problem of extracting geometrical information from a set of eigenvalues of the Laplacian matrix. In the case of 3D-shapes, using FEM, a solution is proposed in [ReuterPeineckeWolter06]

Short Term Developing the idea of image signature Compare different Laplacian definitions Explore the invariance possibilities Experimentation of plugging geometrical information in the learning context

Long Term Wider view for plugging geometrical information in the learning contest Characterize the manifold of images inside the space of matrices Torralba, Fergus and Freeman – 80 million tiny images: a large dataset for non-parametric object and scene recognition -