Some Statistics Stuff (A.K.A. Shamelessly Stolen Stuff)

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Independent Component Analysis
Eigen Decomposition and Singular Value Decomposition
Dimension reduction (2) Projection pursuit ICA NCA Partial Least Squares Blais. “The role of the environment in synaptic plasticity…..” (1998) Liao et.
Color Imaging Analysis of Spatio-chromatic Decorrelation for Colour Image Reconstruction Mark S. Drew and Steven Bergner
Lecture 15 Orthogonal Functions Fourier Series. LGA mean daily temperature time series is there a global warming signal?
Factor Analysis Continued
The General Linear Model Or, What the Hell’s Going on During Estimation?
Dimension reduction (1)
Independent Component Analysis & Blind Source Separation
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Independent Component Analysis (ICA)
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
Dimensional reduction, PCA
Independent Component Analysis & Blind Source Separation Ata Kaban The University of Birmingham.
Principle Component Analysis What is it? Why use it? –Filter on your data –Gain insight on important processes The PCA Machinery –How to do it –Examples.
Face Recognition Jeremy Wyatt.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Clustering In Large Graphs And Matrices Petros Drineas, Alan Frieze, Ravi Kannan, Santosh Vempala, V. Vinay Presented by Eric Anderson.
A Quick Practical Guide to PCA and ICA Ted Brookings, UCSB Physics 11/13/06.
Bayesian belief networks 2. PCA and ICA
Basic Mathematics for Portfolio Management. Statistics Variables x, y, z Constants a, b Observations {x n, y n |n=1,…N} Mean.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Principal Component Analysis Principles and Application.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
Lorelei Howard and Nick Wright MfD 2008
Statistical Methods for long-range forecast By Syunji Takahashi Climate Prediction Division JMA.
Analytical Techniques
Variance and covariance Sums of squares General linear models.
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Objectives of Multiple Regression
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
Measuring Functional Integration: Connectivity Analyses
Presented By Wanchen Lu 2/25/2013
1 1 Slide © 2016 Cengage Learning. All Rights Reserved. The equation that describes how the dependent variable y is related to the independent variables.
Independent Component Analysis on Images Instructor: Dr. Longin Jan Latecki Presented by: Bo Han.
Heart Sound Background Noise Removal Haim Appleboim Biomedical Seminar February 2007.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Analysis of fMRI data with linear models Typical fMRI processing steps Image reconstruction Slice time correction Motion correction Temporal filtering.
Hongyan Li, Huakui Wang, Baojin Xiao College of Information Engineering of Taiyuan University of Technology 8th International Conference on Signal Processing.
Independent Component Analysis Zhen Wei, Li Jin, Yuxue Jin Department of Statistics Stanford University An Introduction.
AN ORTHOGONAL PROJECTION
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Multivariate Statistics Matrix Algebra I W. M. van der Veld University of Amsterdam.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Canonical Correlation Psy 524 Andrew Ainsworth. Matrices Summaries and reconfiguration.
SE-280 Dr. Mark L. Hornick Multiple Regression (Cycle 4)
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
Lecture 3 BME452 Biomedical Signal Processing 2013 (copyright Ali Işın, 2013) 1 BME452 Biomedical Signal Processing Lecture 3  Signal conditioning.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
Feature Selection and Extraction Michael J. Watts
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
Introduction to Independent Component Analysis Math 285 project Fall 2015 Jingmei Lu Xixi Lu 12/10/2015.
CHAPTER- 3.2 ERROR ANALYSIS. 3.3 SPECIFIC ERROR FORMULAS  The expressions of Equations (3.13) and (3.14) were derived for the general relationship of.
Feature Extraction 主講人:虞台文.
Unsupervised Learning II Feature Extraction
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
LECTURE 11: Advanced Discriminant Analysis
Dimension Reduction via PCA (Principal Component Analysis)
Machine Learning Dimensionality Reduction
Bayesian belief networks 2. PCA and ICA
Introduction to Connectivity Analyses
A Fast Fixed-Point Algorithm for Independent Component Analysis
Principal Component Analysis
Presentation transcript:

Some Statistics Stuff (A.K.A. Shamelessly Stolen Stuff)

Blind Source Separation Suppose we have a data set that  Has many independent components or channels  Audio track recorded from multiple microphones  Series of brain images with multiple voxels  Seismographic data from multiple seismometers  We believe is driven by several independent processes  Different people speaking into microphones  Different neuronal processes occurring within the brain  Different parts of the fault moving  We have no a priori notion of what those processes look like Our goal is to figure out what the different processes are by grouping together data that is correlated

ICA (Independent Component Analysis) Put the raw data into a vector X, of length m. We want to find sources or “components”, a vector S, of length n. We imagine that X is produced by some mixing matrix A: X = A * S Or alternatively S is estimated by a weight matrix W: S = W * X We want to find components that are (somehow) maximally independent, yet generate X. * *Note: X isn’t actually raw data

Spelling Things Out The meaning of the ICA equation: we’re choosing a basis. e.g. if W =.6 and W =.2, then S =.6 X +.2 X. That is, X and X are actually being generated (at least partly) by the process S X is typically a time series ---that is, X is measured at discrete intervals. However, our basis doesn’t change, because the fundamental processes that are at work is presumed to be constant. Because of this, W is constant in time, and S changes with time. The end result of ICA is then S(t), and W, which tells us the activity of each component, and how to generate the original data from the components.

Er… “Maximally Independent”? Correlated:Uncorrelated: Technical, and the definition depends somewhat on the algorithm being used. Ultimately boils down to cross-correlations. If two variables are uncorrelated, they are independent. Images from web page by Aapo Hyvärinen,

Maximally Independent is NOT: Orthogonal. In general, AS AS  0. This is different than many other techniques. ICA attempts to group correlated data, but sometimes different parts of the data overlap in a non-orthogonal manner. 1 2.

Requirements  At most one gaussian-distributed element of data  The number of independent data must be greater than the number of components: m > n. E.g. number of microphones greater than number of voices.

For More Info Check out Wikipedia (seriously). The article on independent component analysis is  Actually good.  Provides links to software packages for C++, Java, Matlab, etc. See especially FastICA.  Many of the external links provide good overviews as well.

The Aftermath… Great! Now that we have what we’ve always wanted (a list of “components”) what do we do with them? Since ICA is “blind” it doesn’t tells us much about the components. We may simply be interested in data reduction, or categorizing the mechanisms at work. We may be interested in components that correlate with some signal that we drove the experiment with.

An Alternative or a Supplement: Partial Least Squares Partial Least Squares (PLS) is a technique that is NOT “blind”. PLS attempts to find components that correlate maximally with one or a few signals that drive the system. Example: psychology trials where patients alternate between different tests. Because the result of PLS will be orthogonal components, care must be taken that the signals driving the system are truly orthogonal (or alternatively, they should be transformed to an orthogonal basis before using PLS).

Technical Detail The basic PLS technique is to form a cross-correlation matrix between the signals and the data. Then use singular-value decomposition to find special vectors (analogous to eigenvectors) that explain the variance. These singular vectors are the components, and they are the vectors that minimize the “sum of squares” error when attempting to reconstruct the data from the applied signals.