Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham.

Slides:



Advertisements
Similar presentations
Independent Components Analysis
Advertisements

Independent Component Analysis
Independent Component Analysis: The Fast ICA algorithm
Discovering Cyclic Causal Models by Independent Components Analysis Gustavo Lacerda Peter Spirtes Joseph Ramsey Patrik O. Hoyer.
EE645: Independent Component Analysis
Dimension reduction (2) Projection pursuit ICA NCA Partial Least Squares Blais. “The role of the environment in synaptic plasticity…..” (1998) Liao et.
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Algebraic, transcendental (i.e., involving trigonometric and exponential functions), ordinary differential equations, or partial differential equations...
Dimension reduction (1)
Independent Component Analysis & Blind Source Separation
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
REAL-TIME INDEPENDENT COMPONENT ANALYSIS IMPLEMENTATION AND APPLICATIONS By MARCOS DE AZAMBUJA TURQUETI FERMILAB May RTC 2010.
Independent Component Analysis (ICA)
Application of Statistical Techniques to Neural Data Analysis Aniket Kaloti 03/07/2006.
Dimensional reduction, PCA
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
3/24/2006Lecture notes for Speech Communications Multi-channel speech enhancement Chunjian Li DICOM, Aalborg University.
Subband-based Independent Component Analysis Y. Qi, P.S. Krishnaprasad, and S.A. Shamma ECE Department University of Maryland, College Park.
Independent Component Analysis (ICA) and Factor Analysis (FA)
A Quick Practical Guide to PCA and ICA Ted Brookings, UCSB Physics 11/13/06.
Bayesian belief networks 2. PCA and ICA
Some Statistics Stuff (A.K.A. Shamelessly Stolen Stuff)
Independent Component Analysis From PCA to ICA Bell Sejnowski algorithm Kurtosis method Demonstrations.
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Linear and Non-Linear ICA-BSS I C A  Independent Component Analysis B S S  Blind Source Separation Carlos G. Puntonet Dept.of Architecture.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Multidimensional Data Analysis : the Blind Source Separation problem. Outline : Blind Source Separation Linear mixture model Principal Component Analysis.
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Independent Components Analysis with the JADE algorithm
Independent Component Analysis on Images Instructor: Dr. Longin Jan Latecki Presented by: Bo Han.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
Hongyan Li, Huakui Wang, Baojin Xiao College of Information Engineering of Taiyuan University of Technology 8th International Conference on Signal Processing.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
Blind Source Separation by Independent Components Analysis Professor Dr. Barrie W. Jervis School of Engineering Sheffield Hallam University England
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Image cryptosystems based on PottsNICA algorithms Meng-Hong Chen Jiann-Ming Wu Department of Applied Mathematics National Donghwa University.
Computational Intelligence: Methods and Applications Lecture 8 Projection Pursuit & Independent Component Analysis Włodzisław Duch Dept. of Informatics,
Lecture 2: Statistical learning primer for biologists
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
Independent Component Analysis Independent Component Analysis.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
ICA and PCA 學生:周節 教授:王聖智 教授. Outline Introduction PCA ICA Reference.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Object Orie’d Data Analysis, Last Time
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 09: Discriminant Analysis Objectives: Principal.
HST.582J/6.555J/16.456J Gari D. Clifford Associate Director, Centre for Doctoral Training, IBME, University of Oxford
Unsupervised Learning II Feature Extraction
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
By: Soroosh Mariooryad Advisor: Dr.Sameti 1 BSS & ICA Speech Recognition - Spring 2008.
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
LECTURE 11: Advanced Discriminant Analysis
Brain Electrophysiological Signal Processing: Preprocessing
Statistical Models for Automatic Speech Recognition
PCA vs ICA vs LDA.
Bayesian belief networks 2. PCA and ICA
Statistical Models for Automatic Speech Recognition
A Fast Fixed-Point Algorithm for Independent Component Analysis
Feature space tansformation methods
Isolation Of Individual Tracks From Polyphonic Music
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham

Overview Today we learn about –The cocktail party problem -- called also ‘blind source separation’ (BSS) –Independent Component Analysis (ICA) for solving BSS –Other applications of ICA / BSS At an intuitive & introductory & practical level

A bit like… in the sense of having to find quantities that are not observable directly

Signals, joint density time Amplitude S 1 (t) Amplitude S 2 (t) Signals Joint density marginal densities

Original signals (hidden sources) s 1 (t), s 2 (t), s 3 (t), s 4 (t), t=1:T

The ICA model s1s1 s2s2 s3s3 s4s4 x1x1 x2x2 x3x3 x4x4 a 11 a 12 a 13 a 14 x i (t) = a i1 *s 1 (t) + a i2 *s 2 (t) + a i3 *s 3 (t) + a i4 *s 4 (t) Here, i=1:4. In vector-matrix notation, and dropping index t, this is x = A * s

This is recorded by the microphones: a linear mixture of the sources x i (t) = a i1 *s 1 (t) + a i2 *s 2 (t) + a i3 *s 3 (t) + a i4 *s 4 (t)

The coctail party problem Called also Blind Source Separation (BSS) problem Ill posed problem, unless assumptions are made! The most common assumption is that source signals are statistically independent. This means that knowing the value of one of them does not give any information about the other. The methods based on this assumption are called Independent Component Analysis methods. These are statistical techniques of decomposing a complex data set into independent parts. It can be shown that under some reasonable conditions, if the ICA assumption holds, then the source signals can be recovered up to permutation and scaling. Determine the source signals, given only the mixtures

Recovered signals

Some further considerations If we knew the mixing parameters a ij then we would just need to solve a linear system of equations. We know neither a ij nor s i. ICA was initially developed to deal with problems closely related to the coctail party problem Later it became evident that ICA has many other applications too. E.g. from electrical recordings of brain activity from different locations of the scalp (EEG signals) recover underlying components of brain activity

Illustration of ICA with 2 signals s1s1 s2s2 x1x1 x2x2 Original sMixed signals a2a2 a1a1 a1a1

Illustration of ICA with 2 signals x1x1 x2x2 Step1: Sphering Step2: Rotatation Mixed signals a2a2 a1a1 a1a1

Illustration of ICA with 2 signals s1s1 s2s2 x1x1 x2x2 Step1: Sphering Step2: Rotatation Original sMixed signals a2a2 a1a1 a1a1

Excluded case There is one case when rotation doesn’t matter. This case cannot be solved by basic ICA. …when both densities are Gaussian Example of non-Gaussian density (-) vs.Gaussian (-.) Seek non-Gaussian sources for two reasons: * identifiability * interestingness: Gaussians are not interesting since the superposition of independent sources tends to be Gaussian

Computing the pre-processing steps for ICA 0) Centring = make the signals centred in zero x i  x i - E[x i ] for each i 1) Sphering = make the signals uncorrelated. I.e. apply a transform V to x such that Cov(Vx)=I // where Cov(y)=E[yy T ] denotes covariance matrix V=E[xx T ] -1/2 // can be done using ‘sqrtm’ function in MatLab x  Vx // for all t (indexes t dropped here) // bold lowercase refers to column vector; bold upper to matrix Scope: to make the remaining computations simpler. It is known that independent variables must be uncorrelated – so this can be fulfilled before proceeding to the full ICA

Fixed Point Algorithm Input: X Random init of W Iterate until convergence: Output: W, S where g(.) is derivative of G(.), W is the rotation transform sought Λ is Lagrange multiplier to enforce that W is an orthogonal transform i.e. a rotation Solve by fixed point iterations The effect of Λ is an orthogonal de-correlation Aapo Hyvarinen (97) Computing the rotation step This is based on an the maximisation of an objective function G(.) which contains an approximate non-Gaussianity measure.  The overall transform then to take X back to S is (W T V)  There are several g(.) options, each will work best in special cases. See FastICA sw / tut for details.

Application domains of ICA Blind source separation (Bell&Sejnowski, Te won Lee, Girolami, Hyvarinen, etc.) Image denoising (Hyvarinen) Medical signal processing – fMRI, ECG, EEG (Mackeig) Modelling of the hippocampus and visual cortex (Lorincz, Hyvarinen) Feature extraction, face recognition (Marni Bartlett) Compression, redundancy reduction Watermarking (D Lowe) Clustering (Girolami, Kolenda) Time series analysis (Back, Valpola) Topic extraction (Kolenda, Bingham, Kaban) Scientific Data Mining (Kaban, etc)

Image denoising Wiener filtering ICA filtering Noisy image Original image

Clustering In multi-variate data search for the direction along of which the projection of the data is maximally non-Gaussian = has the most ‘structure’

Blind Separation of Information from Galaxy Spectra

Decomposition using Physical Models Decomposition using ICA

Summing Up Assumption that the data consists of unknown components –Individual signals in a mix –topics in a text corpus –basis-galaxies Trying to solve the inverse problem: –Observing the superposition only –Recover components –Components often give simpler, clearer view of the data

Related resources  Demo and links to further info on ICA.  ICA software in MatLab.  Comprehensive tutorial paper, slightly more technical.