A Fast Fixed-Point Algorithm for Independent Component Analysis

Slides:



Advertisements
Similar presentations
Independent Component Analysis
Advertisements

Independent Component Analysis: The Fast ICA algorithm
EE645: Independent Component Analysis
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Independent Component Analysis & Blind Source Separation
Independent Component Analysis (ICA)
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Dimensional reduction, PCA
Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham.
Projection Pursuit. Projection Pursuit (PP) PCA and FDA are linear, PP may be linear or non-linear. Find interesting “criterion of fit”, or “figure of.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Independent Component Analysis (ICA) and Factor Analysis (FA)
An Introduction to Independent Component Analysis (ICA) 吳育德 陽明大學放射醫學科學研究所 台北榮總整合性腦功能實驗室.
Bayesian belief networks 2. PCA and ICA
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Principal Component Analysis. Philosophy of PCA Introduced by Pearson (1901) and Hotelling (1933) to describe the variation in a set of multivariate data.
Normalised Least Mean-Square Adaptive Filtering
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Summarized by Soo-Jin Kim
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
Independent Component Analysis on Images Instructor: Dr. Longin Jan Latecki Presented by: Bo Han.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
Hongyan Li, Huakui Wang, Baojin Xiao College of Information Engineering of Taiyuan University of Technology 8th International Conference on Signal Processing.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Learning Theory Reza Shadmehr LMS with Newton-Raphson, weighted least squares, choice of loss function.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Unsupervised Learning Motivation: Given a set of training examples with no teacher or critic, why do we learn? Feature extraction Data compression Signal.
Computational Intelligence: Methods and Applications Lecture 8 Projection Pursuit & Independent Component Analysis Włodzisław Duch Dept. of Informatics,
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
Independent Component Analysis Independent Component Analysis.
Feature Selection and Extraction Michael J. Watts
ICA and PCA 學生:周節 教授:王聖智 教授. Outline Introduction PCA ICA Reference.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Data Modeling Patrice Koehl Department of Biological Sciences
Principal Component Analysis
Lectures 15: Principal Component Analysis (PCA) and
Deep Feedforward Networks
Ch 12. Continuous Latent Variables ~ 12
LECTURE 06: MAXIMUM LIKELIHOOD ESTIMATION
LECTURE 11: Advanced Discriminant Analysis
Principle Component Analysis (PCA) Networks (§ 5.8)
Brain Electrophysiological Signal Processing: Preprocessing
Overview of Supervised Learning
PCA vs ICA vs LDA.
Bayesian belief networks 2. PCA and ICA
J.-F. Pâris University of Houston
Aapo Hyvärinen and Ella Bingham
EE513 Audio Signals and Systems
Matrix Algebra and Random Vectors
Biointelligence Laboratory, Seoul National University
Learning Theory Reza Shadmehr
Introduction to Radial Basis Function Networks
Parametric Methods Berlin Chen, 2005 References:
Independent Factor Analysis
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME
Computational Intelligence: Methods and Applications
Restructuring Sparse High Dimensional Data for Effective Retrieval
Presentation transcript:

A Fast Fixed-Point Algorithm for Independent Component Analysis Neural Computation, 9:1483-1492, 1997 A. Hyvarinen, E. Oja Summarized by Seong-woo Chung 2001.6.8

(C) 2001, SNU CSE Biointelligence Lab Introduction Independent Component Analysis (ICA) is to express a set of random variables as linear combinations of statistically independent component variables Two applications of ICA are blind source separation and feature extraction (C) 2001, SNU CSE Biointelligence Lab

(C) 2001, SNU CSE Biointelligence Lab Introduction(2) In the simplest form of ICA Observable m scalar variables v1, v2, …, vm n unknown independent components s1, s2, …, sn n < m v (vector) is linear combinations of s (vector) with an unknown m×n matrix (called the mixing matrix) Can only estimate non-Gaussian independent components (except if just one of the independent components is Gaussian) Defines that the independent components si have unit variance (C) 2001, SNU CSE Biointelligence Lab

(C) 2001, SNU CSE Biointelligence Lab Introduction(3) The problem of estimating the matrix A can be simplified by performing sphering or prewhitening of the data v v is linearly transformed to a vector x = Mv such that its elements xi are mutually uncorrelated and all have unit variance Thus the correlation matrix of x equals unity: B=MA is an orthogonal matrix due to assumptions on the components si (C) 2001, SNU CSE Biointelligence Lab

ICA by Kurtosis Minimization and Maximization ICA use the fourth-order cumulant or kurtosis of the signals, defined for a zero-mean random variable v as For a Gaussian, kurtosis is zero, for densities peaked at zero, positive, and for flatter densities, negative To find w satisfing , following object function have to be minimized or maximized (C) 2001, SNU CSE Biointelligence Lab

ICA by Kurtosis Minimization and Maximization(2) Using gradient rule, The advantage is fast adaptation in a non-stationary environment A resulting trade-off is that the convergence is slow, and depends on a good choice of the learning rate sequence μ(t) (C) 2001, SNU CSE Biointelligence Lab

Fixed-Point Algorithm Using the above derivation, we get the following fixed-point algorithm for ICA Take a random initial vector w(0) of norm 1. Let k=1 Let Divide w(k) by its norm If |w(k)w(k-1)| is not close to 1, let k=k+1 and go back to step 2, Otherwise, output the vector w(k) (C) 2001, SNU CSE Biointelligence Lab

(C) 2001, SNU CSE Biointelligence Lab Application Blind source separation Feature extraction (C) 2001, SNU CSE Biointelligence Lab

Blind source separation <Eight independent components of the EEG data> (C) 2001, SNU CSE Biointelligence Lab

(C) 2001, SNU CSE Biointelligence Lab Feature extraction <Some ICA basis vectors of natural image data> (C) 2001, SNU CSE Biointelligence Lab

Discussion(about the Fast Fixed-Point Analysis) The convergence of the algorithm is very fast There is no learning rate or other adustable parameters Finds the independent components one at a time Both components of negative and positive kurtosis can be found (C) 2001, SNU CSE Biointelligence Lab