Download presentation
Presentation is loading. Please wait.
Published byConstance Owen Modified over 9 years ago
2
“A fast method for Underdetermined Sparse Component Analysis (SCA) based on Iterative Detection- Estimation (IDE)” Arash Ali-Amini 1 Massoud BABAIE-ZADEH 1,2 Christian Jutten 2 1- Sharif University of Technology, Tehran, IRAN 2- Laboratory of Images and Signals (LIS), CNRS, INPG, UJF, Grenoble, FRANCE
3
MaxEnt2006, 8-13 July, Paris, France 2/42 Outline Introduction to Blind Source Separation Geometrical Interpretation Sparse Component Analysis (SCA), underdetermined case Identifying mixing matrix Source restoration Finding sparse solutions of an Underdetermined System of Linear Equations (USLE): Minimum L0 norm Matching Pursuit Minimum L1 norm or Basis Pursuit ( Linear Programming) Iterative Detection-Estimation (IDE) - our method Simulation results Conclusions and Perspectives
4
MaxEnt2006, 8-13 July, Paris, France 3/42 Blind Source Separation (BSS) Source signals s 1, s 2, …, s M Source vector: s=(s 1, s 2, …, s M ) T Observation vector: x=(x 1, x 2, …, x N ) T Mixing system x = As Goal Finding a separating matrix y = Bx
5
MaxEnt2006, 8-13 July, Paris, France 4/42 Blind Source Separation (cont.) Assumption: N >=M (#sensors >= #sources) A is full-rank (invertible) prior information: Statistical “Independence” of sources Main idea: Find “B” to obtain “independent” outputs ( Independent Component Analysis=ICA)
6
MaxEnt2006, 8-13 July, Paris, France 5/42 Blind Source Separation (cont.) Separability Theorem [Comon 1994,Darmois 1953]: If at most 1 source is Gaussian: statistical independence of outputs source separation ( ICA: a method for BSS) Indeterminacies: permutation, scale A = [a 1, a 2, …, a M ], x=As x = s 1 a 1 + s 2 a 2 + … + s M a M
7
MaxEnt2006, 8-13 July, Paris, France 6/42 Geometrical Interpretation x=As Statistical Independence of s1 and s2, bounded PDF rectangular support for joint PDF of (s1,s2) rectangular scatter plot for (s1,s2)
8
MaxEnt2006, 8-13 July, Paris, France 7/42 Sparse Sources s-planex-plane 2 sources, 2 sensors:
9
MaxEnt2006, 8-13 July, Paris, France 8/42 Importance of Sparse Source Separation The sources may be not sparse in time, but sparse in another domain (frequency, time- frequency, time-scale, etc.) x=As T{x}=A T{s} T: a ‘sparsifying’ transform Example. Speech signals are usually sparse in ‘time-frequency’ domain.
10
MaxEnt2006, 8-13 July, Paris, France 9/42 Sparse sources (cont.) 3 sparse sources, 2 sensors Sparsity Source Separation, with more sensors than sources?
11
MaxEnt2006, 8-13 July, Paris, France 10/42 Estimating the mixing matrix A = [a 1, a 2, a 3 ] x = s 1 a 1 + s 2 a 2 + s 3 a 3 Mixing matrix is easily identified for sparse sources Scale & Permutation indeterminacy ||a i ||=1
12
MaxEnt2006, 8-13 July, Paris, France 11/42 Restoration of the sources How to find the sources, after having found the mixing matrix (A)? 2 equations, 3 unknowns infinitely many solutions! Underdertermined SCA, underdetermined system of equations
13
MaxEnt2006, 8-13 July, Paris, France 12/42 Identification vs Separation #Sources <= #Sensors: Identifying A Source Separation #Sources > #Sensors (underdetermined) Identifying A Source Separation Two different problems: Identifying the mixing matrix (relatively easy) Restoring the sources (difficult)
14
MaxEnt2006, 8-13 July, Paris, France 13/42 Is it possible, at all? A is known, at eash instant (n 0 ), we should solve un underdetermined linear system of equations: Infinite number of solutions s(n 0 ) Is it possible to recover the sources?
15
MaxEnt2006, 8-13 July, Paris, France 14/42 ‘Sparse’ solution s i (n) sparse in time The vector s(n 0 ) is most likely a ‘sparse vector’ A.s(n 0 ) = x(n 0 ) has infinitely many solutions, but not all of them are sparse! Idea: For restoring the sources, take the sparsest solution (most likely solution)
16
MaxEnt2006, 8-13 July, Paris, France 15/42 Example ( 2 equations, 4 unknowns ) Some of solutions: Sparsest
17
MaxEnt2006, 8-13 July, Paris, France 16/42 The idea of solving underdetermined SCA A s(n) = x(n), n=0,1,…,T Step 1 (identification): Estimate A (relatively easy) Step 2 (source restoration): At each instant n 0, find the sparsest solution of A s( n 0 ) = x( n 0 ), n 0 =0,…,T Main question: HOW to find the sparsest solution of an Underdetermined System of Linear Equations (USLE)?
18
MaxEnt2006, 8-13 July, Paris, France 17/42 Another application of USLE: Atomic decomposition over an overcompelete dictionary Decomposing a signal x, as a linear combination of a set of fixed signals (atoms) Terminology: Atoms: i, i=1,…,M Dictionary: { 1, 2,…, M }
19
MaxEnt2006, 8-13 July, Paris, France 18/42 Atomic decomposition (cont.) M=N Complete dictionary Unique set of coefficients Examples: Dirac dictionary, Fourier Dictionary Dirac Dictionary:
20
MaxEnt2006, 8-13 July, Paris, France 19/42 Atomic decomposition (cont.) M=N Complete dictionary Unique set of coefficients Examples: Dirac dictionary, Fourier Dictionary Fourier Dictionary:
21
MaxEnt2006, 8-13 July, Paris, France 20/42 Atomic decomposition (cont.) Matrix Form: If just a few number of coefficient are non-zero The underlying structure is very well revealed Example. signal has just a few non-zero samples in time its decomposition over the Dirac dictionary reveals it Signals composed of a few pure frequencies its decomposition over the Fourier dictionary reveals it How about a signals which is the sum of a pure frequency and a dirac?
22
MaxEnt2006, 8-13 July, Paris, France 21/42 Atomic decomposition (cont.) Solution: consider a larger dictionary, containing both Dirac and Fourier atoms M>N Overcomplete dictionary. Problem: Non-uniqueness of (Non-unique representation) ( USLE) However: we are looking for sparse solution
23
MaxEnt2006, 8-13 July, Paris, France 22/42 Sparse solution of USLE Underdetermined SCA Atomic Decomposition on over-complete dictionaries Findind sparsest solution of USLE
24
MaxEnt2006, 8-13 July, Paris, France 23/42 Uniqueness of sparse solution x=As, n equations, m unknowns, m>n Question: Is the sparse solution unique? Theorem (Donoho 2004): if there is a solution s with less than n/2 non-zero components, then it is unique with probability 1 (that is, for almost all A’s).
25
MaxEnt2006, 8-13 July, Paris, France 24/42 How to find the sparsest solution A.s = x, n equations, m unknowns, m>n Goal: Finding the sparsest solution Note: at least m-n sources are zero. Direct method: Set m-n (arbitrary) sources equal to zero Solve the remaining system of n equations and n unknowns Do above for all possible choices, and take sparsest answer. Another name: Minimum L 0 norm method L 0 norm of s = number of non-zero components = |s i | 0
26
MaxEnt2006, 8-13 July, Paris, France 25/42 Example s1=s2=0 s=(0, 0, 1.5, 2.5) T L 0 =2 s1=s3=0 s=(0, 2, 0, 0) T L 0 =1 s1=s4=0 s=(0, 2, 0, 0) T L 0 =1 s2=s3=0 s=(2, 0, 0, 2) T L 0 =2 s2=s4=0 s=(10, 0, -6, 0) T L 0 =2 s3=s4=0 s=(0, 2, 0, 0) T L 0 =2 Minimum L 0 norm solution s=(0, 2, 0, 0) T
27
MaxEnt2006, 8-13 July, Paris, France 26/42 Drawbacks of minimal norm L 0 Highly (unacceptably) sensitive to noise Need for a combinatorial search: Example. m=50, n=30, On our computer: Time for solving a 30 by 30 system of equation=2x10 -4 s Total time (5x10 13 )(2x10 -4 ) 300 years! Non-tractable
28
MaxEnt2006, 8-13 July, Paris, France 27/42 faster methods? Matching Pursuit [Mallat & Zhang, 1993] Basis Pursuit (minimal L1 norm Linear Programming) [Chen, Donoho, Saunders, 1995] Our method (IDE)
29
MaxEnt2006, 8-13 July, Paris, France 28/42 Matching Pursuit (MP) [Mallat & Zhang, 1993]
30
MaxEnt2006, 8-13 July, Paris, France 29/42 Properties of MP Advantage: Very Fast Drawback A very ‘greedy’ algorithm Mistake in a stage can never be corrected Not necessarily a sparse solution x=s1a1+s2a2x=s1a1+s2a2 a1 a2 a3 a5 a4
31
MaxEnt2006, 8-13 July, Paris, France 30/42 Minimum L 1 norm or Basis Pursuit [Chen, Donoho, Saunders, 1995] Minimum norm L1 solution: MAP estimator under a Laplacian prior Recent theoretical support (Donoho, 2004) : For ‘most’ ‘large’ underdetermined systems of linear equations, the minimal L 1 norm solution is also the sparsest solution
32
MaxEnt2006, 8-13 July, Paris, France 31/42 Minimal L 1 norm (cont.) Minimal L 1 norm solution may be found by Linear Programming (LP) Fast algorithms for LP: Simplex Interior Point method
33
MaxEnt2006, 8-13 July, Paris, France 32/42 Minimal L 1 norm (cont.) Advantages: Very good practical results Theoretical support Drawback: Tractable, but still very time-consuming
34
MaxEnt2006, 8-13 July, Paris, France 33/42 Iterative Detection-Estmation (IDE)- Our method Main Idea: Step 1 (Detection): Detect which sources are ‘active’, and which are ‘non-active’ Step 2 (Estimation): Knowing active sources, estimate their values Problem: Detecting the activity status of a source, requires the values of all other sources! Our proposition: Iterative Detection-Estimation Activity Detection Value Estimation
35
MaxEnt2006, 8-13 July, Paris, France 34/42 IDE, Detection step Assumptions: Mixture of Gaussian model for sparse sources: s i active s i ~N(0, 1 2 ) s i in-active s i ~N(0, 0 2 ), 1 >> 0 0 : probability of in-activity 1 Columns of A are normalized Hypotheses: Proposition: t=a 1 T x=s 1 + Activity detection test: g 1 (s)=|t- |> depends on other sources use their previous estimates
36
MaxEnt2006, 8-13 July, Paris, France 35/42 IDE, Estimation step Estimation equation: Solution using Lagrange multipliers: Closed-form solution See the paper
37
MaxEnt2006, 8-13 July, Paris, France 36/42 IDE, summary Detection Step (resulted from binary hypothesis testing, with a Mixture of Gaussian source model): Estimation Step:
38
MaxEnt2006, 8-13 July, Paris, France 37/42 IDE, simulation results m=1024, n=0.4m=409
39
MaxEnt2006, 8-13 July, Paris, France 38/42 IDE, simulation results (cont.) m=1024, n=0.4m=409 IDE is about two order of magnitudes faster than LP method (with interior point optimization).
40
MaxEnt2006, 8-13 July, Paris, France 39/42 Speed/Complexity comparision
41
MaxEnt2006, 8-13 July, Paris, France 40/42 IDE, simulation results: number of possible active sources m=500, n=0.4m=200, averaged on 10 simulations
42
MaxEnt2006, 8-13 July, Paris, France 41/42 Conclusion and Perspectives Two problems of Underdetermined SCA: Identifying mixing matrix Restoring sources Two applications of finding sparse solution of USLE’s: Source restoration in underdetermined SCA Atomic Decomposition on over-complete dictionaries Methods: Minimum L0 norm ( Combinatorial search) Minimum L1 norm or Basis Pursuit ( Linear Programming) Matching Pursuit Iterative Detection-Estimation (IDE) Perspectives: Better activity detection (removing thresholds?) Applications in other domains
43
MaxEnt2006, 8-13 July, Paris, France 42/42 Thank you very much for your attention
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.