Presentation is loading. Please wait.

Presentation is loading. Please wait.

Dimensionality reduction

Similar presentations


Presentation on theme: "Dimensionality reduction"— Presentation transcript:

1 Dimensionality reduction
Usman Roshan

2 Dimensionality reduction
What is dimensionality reduction? Compress high dimensional data into lower dimensions How do we achieve this? PCA (unsupervised): We find a vector w of length 1 such that the variance of the projected data onto w is maximized. Binary classification (supervised): Find a vector w that maximizes ratio (Fisher) or difference (MMC) of means and variances of the two classes.

3 PCA Find projection that maximizes variance
PCA minimizes reconstruction error How many dimensions to reduce data to? Consider difference between consecutive eigenvalues If it exceeds a threshold we stop there.s

4 Feature extraction vs selection
PCA and other dimensionality reduction algorithms (to follow) allow feature extraction and selection. In extraction we consider a linear combination of all features. In selection we pick specific features from the data.

5 Kernel PCA Main idea of kernel version
XXTw = λw XTXXTw = λXTw (XTX)XTw = λXTw XTw is projection of data on the eigenvector w and also the eigenvector of XTX This is also another way to compute projections in space quadratic in number of rows but only gives projections.

6 Kernel PCA In feature space the mean is given by
Suppose for a moment that the data is mean subtracted in feature space. In other words mean is 0. Then the scatter matrix in feature space is given by

7 Kernel PCA The eigenvectors of ΣΦ give us the PCA solution. But what if we only know the kernel matrix? First we center the kernel matrix so that mean is 0 where j is a vector of 1’s.K = K

8 Kernel PCA Recall from earlier Same idea for kernel PCA
XXTw = λw XTXXTw = λXTw (XTX)XTw = λXTw XTw is projection of data on the eigenvector w and also the eigenvector of XTX XTX is the linear kernel matrix Same idea for kernel PCA The projected solution is given by the eigenvectors of the centered kernel matrix.

9 Polynomial degree 2 kernel Breast cancer

10 Polynomial degree 2 kernel Climate

11 Polynomial degree 2 kernel Qsar

12 Polynomial degree 2 kernel Ionosphere

13 Supervised dim reduction: Linear discriminant analysis
Fisher linear discriminant: Maximize ratio of difference means to sum of variance

14 Linear discriminant analysis
Fisher linear discriminant: Difference in means of projected data gives us the between-class scatter matrix Variance gives us within-class scatter matrix

15 Linear discriminant analysis
Fisher linear discriminant solution: Take derivative w.r.t. w and set to 0 This gives us w = cSw-1(m1-m2)

16 Scatter matrices Sb is between class scatter matrix
Sw is within-class scatter matrix St = Sb + Sw is total scatter matrix

17 Fisher linear discriminant
General solution is given by eigenvectors of Sb-1Sw

18 Fisher linear discriminant
Problems can happen with calculating the inverse A different approach is the maximum margin criterion

19 Maximum margin criterion (MMC)
Define the separation between two classes as S(C) represents the variance of the class. In MMC we use the trace of the scatter matrix to represent the variance. The scatter matrix is

20 Maximum margin criterion (MMC)
The scatter matrix is The trace (sum of diagonals) is Consider an example with two vectors x and y

21 Maximum margin criterion (MMC)
Plug in trace for S(C) and we get The above can be rewritten as Where Sw is the within-class scatter matrix And Sb is the between-class scatter matrix

22 Weighted maximum margin criterion (WMMC)
Adding a weight parameter gives us In WMMC dimensionality reduction we want to find w that maximizes the above quantity in the projected space. The solution w is given by the largest eigenvector of the above

23 How to use WMMC for classification?
Reduce dimensionality to fewer features Run any classification algorithm like nearest means or nearest neighbor.

24 K-nearest neighbor Classify a given datapoint to be the majority label of the k closest points The parameter k is cross-validated Simple yet can obtain high classification accuracy

25 Weighted maximum variance (WMV)
Find w that maximizes the weighted variance

26 Weighted maximum variance (WMV)
Reduces to PCA if Cij = 1/n


Download ppt "Dimensionality reduction"

Similar presentations


Ads by Google