A principled way to principal components analysis

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Face Recognition Sumitha Balasuriya.
Dimensionality Reduction PCA -- SVD
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
Principal Component Analysis
3D Geometry for Computer Graphics
Unsupervised Learning - PCA The neural approach->PCA; SVD; kernel PCA Hertz chapter 8 Presentation based on Touretzky + various additions.
Self Organization: Hebbian Learning CS/CMPE 333 – Neural Networks.
Un Supervised Learning & Self Organizing Maps Learning From Examples
Face Recognition Jeremy Wyatt.
Perceptron Learning Rule Assuming the problem is linearly separable, there is a learning rule that converges in a finite time Motivation A new (unseen)
Lecture 4 Unsupervised Learning Clustering & Dimensionality Reduction
Face Recognition Using Eigenfaces
SVD and PCA COS 323. Dimensionality Reduction Map points in high-dimensional space to lower number of dimensionsMap points in high-dimensional space to.
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Microarray analysis Algorithms in Computational Biology Spring 2006 Written by Itai Sharon.
A globally asymptotically stable plasticity rule for firing rate homeostasis Prashant Joshi & Jochen Triesch
Bayesian belief networks 2. PCA and ICA
3D Geometry for Computer Graphics
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 9 Data Analysis Martin Russell.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Unsupervised learning
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Chapter 2 Dimensionality Reduction. Linear Methods
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Unsupervised learning
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
CSE 185 Introduction to Computer Vision Face Recognition.
CSSE463: Image Recognition Day 27 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Contents PCA GHA APEX Kernel PCA CS 476: Networks of Neural Computation, CSD, UOC, 2009 Conclusions WK9 – Principle Component Analysis CS 476: Networks.
CpSc 881: Machine Learning PCA and MDS. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various sources.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principal Component Analysis Zelin Jia Shengbin Lin 10/20/2015.
Irena Váňová. B A1A1. A2A2. A3A3. repeat until no sample is misclassified … labels of classes Perceptron algorithm for i=1...N if then end * * * * *
Presented by: Muhammad Wasif Laeeq (BSIT07-1) Muhammad Aatif Aneeq (BSIT07-15) Shah Rukh (BSIT07-22) Mudasir Abbas (BSIT07-34) Ahmad Mushtaq (BSIT07-45)
CSSE463: Image Recognition Day 25 This week This week Today: Applications of PCA Today: Applications of PCA Sunday night: project plans and prelim work.
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Principal Component Analysis (PCA)
CSSE463: Image Recognition Day 27
Developing Computational Skills in the Sciences with MATLAB
CSSE463: Image Recognition Day 26
Dimensionality Reduction
Principle Component Analysis (PCA) Networks (§ 5.8)
University of Ioannina
PRINCIPAL COMPONENT ANALYSIS (PCA)
Lecture 8:Eigenfaces and Shared Features
Lecture: Face Recognition and Feature Reduction
Unsupervised Learning: Principle Component Analysis
Principal Component Analysis (PCA)
Machine Learning Dimensionality Reduction
Bayesian belief networks 2. PCA and ICA
Covariation Learning and Auto-Associative Memory
Principal Component Analysis
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Recitation: SVD and dimensionality reduction
X.1 Principal component analysis
CSSE463: Image Recognition Day 25
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
CSSE463: Image Recognition Day 25
Sparseness and Expansion in Sensory Representations
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Dimensionality reduction in neuroscience
Presentation transcript:

A principled way to principal components analysis Daniel Zysman Lecturer

Teaching activity objectives Visualize large data sets. Transform the data to aid in this visualization. Clustering data. Implement basic linear algebra operations. Connect this operations to neuronal models and brain function.

Context for the activity Homework Assignment in 9.40 Intro to neural Computation (Sophomore/Junior). In-class activity 9.014 Quantitative Methods and Computational Models in Neuroscience (1st year PhD).

Data visualization and performing pca:

MNIST data set 28 by 28 pixels 8-bit gray scale images These images live in a 784 dimensional space http://yann.lecun.com/exdb/mnist/

Can we cluster images in the pixel space?

One possible visualization There are more than 300000 possible pairwise pixel plots!!!

Is there a more principled way? Represent the data in a new basis set. Aids in visualization and potentially in clustering and dimensionality reduction. PCA provides such a basis set by looking at directions that capture most variance. The directions are ranked by decreasing variance. It diagonalizes the covariance matrix.

Pedagogical approach Guide them step by step to implement PCA. Emphasize visualizations and geometrical approach/intuition. We don’t use the MATLAB canned function for PCA. We want students to get their hands “dirty”. This helps build confidence and deep understanding.

PCA Mantra Reshape the data to proper format for PCA. Center the data performing mean subtraction. Construct the data covariance matrix. Perform SVD to obtain the eigenvalues and eigenvectors of the covariance matrix. Compute the variance explained per component and plot it. Reshape the eigenvectors and visualize their images. Project the mean subtracted data onto the eigenvectors basis.

First 9 Eigenvectors

Projections onto the first 2 axes The first two PCs capture ~37% of the variance. The data forms clear clusters that are almost linearly separable

Building models: Synapses and PCA

Hebbian Learning 1949 book: 'The Organization of Behavior' Theory about the neural bases of learning Learning takes place at synapses. Synapses get modified, they get stronger when the pre- and post- synaptic cells fire together. "Cells that fire together, wire together" Donald Hebb

Building Hebbian synapses Unstable

Oja’s rule Stabilizes the Hebbian rule. Erkki Oja Feedback,forgetting term or regularizer Stabilizes the Hebbian rule. Leads to a covariance learning rule: the weights converge to the first eigenvector of the covariance matrix. Similar to power iteration method. A simplified neuron model as a principal component analyzer. Journal of Mathematical Biology, 15:267-273 (1982).

Learning outcomes Visualize and manipulate a relatively large and complex data set. Perform PCA by building it step by step. Gain an intuition of the geometry involved in a change of basis and projections. Start thinking about basic clustering algorithms. Discuss on dimensionality reduction and other PCA applications

Learning outcomes (cont) Discuss the assumptions, limitations and shortcomings of applying PCA in different contexts. Build a model of how PCA might actually take place in neural circuits. Follow up: eigenfaces, is the brain doing PCA to recognize faces?