Explaining High-Dimensional Data

Slides:



Advertisements
Similar presentations
EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
Advertisements

Linear Subspaces - Geometry. No Invariants, so Capture Variation Each image = a pt. in a high-dimensional space. –Image: Each pixel a dimension. –Point.
3D Geometry for Computer Graphics
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Mutidimensional Data Analysis Growth of big databases requires important data processing.  Need for having methods allowing to extract this information.
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Eigenvalues and Eigenvectors
3D Geometry for Computer Graphics
Computer Graphics Recitation 6. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
3D Geometry for Computer Graphics
Proximal Support Vector Machine Classifiers KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Data Mining Institute University of.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Face Recognition Jeremy Wyatt.
TFIDF-space  An obvious way to combine TF-IDF: the coordinate of document in axis is given by  General form of consists of three parts: Local weight.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
3D Geometry for Computer Graphics
Bayesian belief networks 2. PCA and ICA
Olga Sorkine’s slides Tel Aviv University. 2 Spectra and diagonalization A If A is symmetric, the eigenvectors are orthogonal (and there’s always an eigenbasis).
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
The Multivariate Normal Distribution, Part 1 BMTRY 726 1/10/2014.
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 6. Eigenvalue problems.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
SVD(Singular Value Decomposition) and Its Applications
Summarized by Soo-Jin Kim
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
REVIEW OF MATHEMATICS. Review of Vectors Analysis GivenMagnitude of vector: Example: Dot product:  is the angle between the two vectors. Example:
+ Review of Linear Algebra Optimization 1/14/10 Recitation Sivaraman Balakrishnan.
Machine Vision for Robots
Principles of Pattern Recognition
Camera Geometry and Calibration Thanks to Martial Hebert.
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Whitening.
Domain Range definition: T is a linear transformation, EIGENVECTOR EIGENVALUE.
CSE 185 Introduction to Computer Vision Face Recognition.
3D Game Engine Design 1 3D Game Engine Design Ch D MAP LAB.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
KEY THEOREMS KEY IDEASKEY ALGORITHMS LINKED TO EXAMPLES next.
5.1 Eigenvectors and Eigenvalues 5. Eigenvalues and Eigenvectors.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
Introduction to Optimization
1 Overview representing region in 2 ways in terms of its external characteristics (its boundary)  focus on shape characteristics in terms of its internal.
3D Geometry for Computer Graphics Class 3. 2 Last week - eigendecomposition A We want to learn how the transformation A works:
5 5.1 © 2016 Pearson Education, Ltd. Eigenvalues and Eigenvectors EIGENVECTORS AND EIGENVALUES.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Unsupervised Learning II Feature Extraction
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Reduced echelon form Matrix equations Null space Range Determinant Invertibility Similar matrices Eigenvalues Eigenvectors Diagonabilty Power.
SUPPORT VECTOR MACHINES
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Eigenvalues and Eigenvectors
University of Ioannina
LECTURE 10: DISCRIMINANT ANALYSIS
Core-Sets and Geometric Optimization problems.
Lecture 8:Eigenfaces and Shared Features
Probabilistic Models with Latent Variables
Bayesian belief networks 2. PCA and ICA
Principal Component Analysis
Eigenvalues and Eigenvectors
Feature space tansformation methods
Bayesian Classification
LECTURE 09: DISCRIMINANT ANALYSIS
Part 3. Linear Programming
Eigenvalues and Eigenvectors
Eigenvalues and Eigenvectors
Presentation transcript:

Explaining High-Dimensional Data Hoa Nguyen, Rutgers University Mentors: Ofer Melnik Kobbi Nissim

High-Dimensional Data A great deal of data from different domains (medicine, finance, science) is high-dimensional. High-dimensional data is hard to visualize and understand.

Example: Given a set of images (represented by 8x8 pixel matrices), we could consider each image as a point in 64-dimensional space.

Analyzing Data Instead of visualizing, we find properties of data to describe it. Statistics, Data Mining, Machine Learning: Finding properties of data directly Finding models to capture data

Classifier Given a set of data points with assigned labels. Build a model for the data. Use this model to label new unclassified data points.

The Geometrical View of Classifiers: Example: Given a set of data points in 2 dimensional-space with a + or – label for each point: (x i ,y i)± _ _ _ _ + + + _ + + + + _ _ _ We are interested in classifiers that take all the points of a class, and enclose them in a convex shape.

Goal To understand the properties of the data enclosed by a convex shape.

Convex Shapes The convex hull: Problem: Our solution: The convex hull C of a set of points is the smallest convex set that includes all the points. Problem: It is difficult to study the convex hull directly. Our solution: Instead of looking at the convex hull, we use the simpler convex shape to approximate the hull – the ellipsoid.

MVE (Minimum Volume Ellipsoid) Example: + + + + + + + + + + Advantage: Use an ellipsoid to approximate the convex region, or to bound the geometry of the convex hull.

Using MVE to approximate the convex hull [John 1948] has shown that if we shrink the minimum volume outer ellipsoid of a convex set C by a factor k about its center, we obtain an ellipsoid contained in C. (k is the dimension of the space) Example: + + + + + + + + + +

Calculate the MVE The MVE is described by the equation: v’ -1 v = k V={v1,v2,…,vh}  Rk. v is an exterior point : the scatter matrix  = wiviviT the eigen vectors of  correspond to the directions of the ellipsoid axes. the eigen values of correspond to the half-lengths or radii of the axes. k: a constant equal to the dimension of the space wi: the weight of a point vi. h i=1

Calculate the MVE (cont.) [Titterington 1978] An algorithm to calculate the weights of MVE: A point has a positive weight if it lies on the surface of the ellipsoid. At least k+1 points have non-zero weights. At most k(k+3)/2+1 points have non-zero weights.

Use MVE for data analysis: Finding extreme points by looking at points on the ellipsoid surface. Finding the subspace of data by looking at the directions where the ellipsoid is thin.

Points on the surface of the ellipsoid i.e., points with non-zero weights. Example: In our hand-written digit file, there are 376 points which belong to class “0”. By using MVE, we find that there are about 178 points with non-zero weights, i.e., these points lie on the surface of the MVE. The mean-zero Some “0” points on the surface of the MVE

Directions where the ellipsoid is thin The directions and size of an ellipsoid’s axes correspond to the eigen vectors and values of its scatter matrix. Direction of thinness: A short axis defines a direction in which the data does not extend. If V is a zero-valued eigen vector, then it defines a constraint for any data point x: Vx=0

A simple Null Space Any basis for the Null Space of the scatter matrix is an equivalent set of constraints. In order to understand the data, we would like to find constraints that are easy to interpret. Goal: simplify the null space basis, e.g.,find a basis with many zeros.

The Null Space Problem (NSP) The Null Space Problem is defined as finding the basis with the maximal number of zeros. It is an NP-hard problem. [Pothen, Coleman 1986] An approach: Find a heuristic algorithm to simplify an existing basis of the null space of Class “0”: e.g., using Gaussian elimination to get a null space basis with more 0 components.

The null space basis = the set of eigenvectors with 0-eigenvalues The null space basis after using Gaussian elimination

Summary Data analysis Classifier with convex shape MVE Points on the surface of the MVE Simple basis for the null space