Download presentation
Presentation is loading. Please wait.
Published bySusanne Waldfogel Modified over 5 years ago
1
Sparse Modeling of Data and its Relation to Deep Learning
Michael Elad Computer Science Department The Technion - Israel Institute of Technology Haifa 32000, Israel 27-28 June 2019 Series of Lectures on Waves and Imaging (III)
2
A Proposed Theory for Deep-Learning (DL)
This Lecture is About … A Proposed Theory for Deep-Learning (DL) Explanation: DL has been extremely successful in solving a variety of learning problems DL is an empirical field, with numerous tricks and know-how, but almost no theoretical foundations A theory for DL has become the holy-grail of current research in Machine-Learning and related fields Michael Elad The Computer-Science Department The Technion
3
Who Needs Theory ? We All Do !!
“Machine learning has become alchemy” Ali Rahimi: NIPS 2017 Test-of-Time Award We All Do !! … because … A theory … could bring the next rounds of ideas to this field, breaking existing barriers and opening new opportunities … could map clearly the limitations of existing DL solutions, and point to key features that control their performance … could remove the feeling with many of us that DL is a “dark magic”, turning it into a solid scientific discipline Yan LeCun Understanding is a good thing … but another goal is inventing methods. In the history of science and technology, engineering preceded theoretical understanding: Lens & telescope Optics Steam engine Thermodynamics Airplane Aerodynamics Radio & Comm. Info. Theory Computer Computer Science Michael Elad The Computer-Science Department The Technion
4
A Theory for DL ? Architecture Data Algorithms
Raja Giryes (TAU): Studied the architecture of DNN in the context of their ability to give distance-preserving embedding of signals Gitta Kutyniok (TU) & Helmut Bolcskei (ETH): Studied the ability of DNN architectures to approximate families of functions Stephane Mallat (ENS) & Joan Bruna (NYU): Proposed the scattering transform (wavelet-based) and emphasized the treatment of invariances in the input data Richard Baraniuk & Ankit Patel (RICE): Offered a generative probabilistic model for the data, showing how classic architectures and learning algorithms relate to it Data Architecture Algorithms Rene Vidal (JHU): Explained the ability to optimize the typical non- convex objective and yet get to a global minima Naftali Tishby (HUJI): Introduced the Information Bottleneck (IB) concept and demonstrated its relevance to deep learning Stefano Soatto’s team (UCLA): Analyzed the Stochastic Gradient Descent (SGD) algorithm, connecting it to the IB objective Michael Elad The Computer-Science Department The Technion
5
none of them is complete
So, is there a Theory for DL ? The answer is tricky: There are already various such attempts, and some of them are truly impressive … but … none of them is complete Michael Elad The Computer-Science Department The Technion
6
Interesting Observations
Theory origins: Signal Proc., Control Theory, Info. Theory, Harmonic Analysis, Sparse Represen., Quantum Physics, PDE, Machine learning … Ron Kimmel: “DL is a dark monster covered with mirrors. Everyone sees his reflection in it …” David Donoho: “… these mirrors are taken from Cinderella's story, telling each that he is the most beautiful” Today’s talk is on our proposed theory: Architecture Algorithms Data Vardan Papyan Yaniv Romano Jeremias Sulam Aviad Aberdam … and our theory is the best Michael Elad The Computer-Science Department The Technion
7
This Lecture: More Specifically
Sparseland Sparse Representation Theory CSC Convolutional Sparse Coding ML-CSC Multi-Layered Convolutional Sparse Coding Sparsity-Inspired Models Deep-Learning Another underlying idea that accompanies us Generative modeling of data sources enables A systematic algorithm development, & A theoretical analysis of their performance Disclaimer: Being a lecture on the theory of DL, this lecture is … theoretical … and mathematically oriented Michael Elad The Computer-Science Department The Technion
8
Multi-Layered Convolutional Sparse Modeling
Our eventual goal in today’s talk is to present the … So, lets use this as our running title, parse it into words, and explain each of them Michael Elad The Computer-Science Department The Technion
9
Multi-Layered Convolutional Sparse Modeling
Michael Elad The Computer-Science Department The Technion
10
Using models Our Data is Structured
Seismic Data Matrix Data Text Documents Stock Market Biological Signals Videos Radar Imaging Still Images Social Networks Traffic info Using models We are surrounded by various diverse sources of massive information Each of these sources have an internal structure, which can be exploited This structure, when identified, is the engine behind the ability to process data How to identify structure? Voice Signals 3D Objects Medical Imaging Michael Elad The Computer-Science Department The Technion
11
Models A model: a mathematical description of the underlying signal of interest, describing our beliefs regarding its structure The following is a partial list of commonly used models for images Good models should be simple while matching the signals Models are almost always imperfect Principal-Component-Analysis Gaussian-Mixture Markov Random Field Laplacian Smoothness DCT concentration Wavelet Sparsity Piece-Wise-Smoothness C2-smoothness Besov-Spaces Total-Variation Beltrami-Flow Simplicity Reliability Michael Elad The Computer-Science Department The Technion
12
Data Models and Their Use
What this Talk is all About? Data Models and Their Use Almost any task in data processing requires a model – true for denoising, deblurring, super-resolution, inpainting, compression, anomaly-detection, sampling, recognition, separation, and more Sparse and Redundant Representations offer a new and highly effective model – we call it Sparseland We shall describe this and descendant versions of it that lead all the way to … deep-learning Michael Elad The Computer-Science Department The Technion
13
Multi-Layered Convolutional Sparse Modeling
Michael Elad The Computer-Science Department The Technion
14
Sparseland A New Emerging Model Synthesis Recognition Clustering
Mathematics Machine Learning Signal Processing Wavelet Theory Signal Transforms Multi-Scale Analysis Approximation Theory Linear Algebra Optimization Theory Sparseland Interpolation Segmentation Semi-Supervised Learning Sensor-Fusion Source-Separation Inference (solving inverse problems) Classification Summarizing Compression Prediction Denoising Anomaly detection Synthesis Recognition Clustering Identification Michael Elad The Computer-Science Department The Technion
15
The Sparseland Model Task: model image patches of size 8×8 pixels Σ
We assume that a dictionary of such image patches is given, containing 256 atom images The Sparseland model assumption: every image patch can be described as a linear combination of few atoms α1 α2 α3 Σ Michael Elad The Computer-Science Department The Technion
16
The Sparseland Model Properties of this model: Sparsity and Redundancy
α1 α2 α3 Σ We start with a 8-by-8 pixels patch and represent it using 256 numbers – This is a redundant representation However, out of those 256 elements in the representation, only 3 are non-zeros – This is a sparse representation Bottom line in this case: 64 numbers representing the patch are replaced by 6 (3 for the indices of the non-zeros, and 3 for their entries) Michael Elad The Computer-Science Department The Technion
17
Chemistry of Data We could refer to the Sparseland model as the chemistry of information: α1 α2 α3 Σ Our dictionary stands for the Periodic Table containing all the elements Our model follows a similar rationale: Every molecule is built of few elements Michael Elad The Computer-Science Department The Technion
18
M Sparseland : A Formal Description x 𝐃 α
A Dictionary Every column in 𝐃 (dictionary) is a prototype signal (atom) M The vector is generated with few non-zeros at arbitrary locations and values A sparse vector n x 𝐃 α This is a generative model that describes how (we believe) signals are created Michael Elad The Computer-Science Department The Technion
19
Difficulties with Sparseland
Problem 1: Given a signal, how can we find its atom decomposition? A simple example: There are 2000 atoms in the dictionary The signal is known to be built of 15 atoms possibilities If each of these takes 1nano-sec to test, this will take ~7.5e20 years to finish !!!!!! So, are we stuck? α1 α2 α3 Σ Michael Elad The Computer-Science Department The Technion
20
Approximation Algorithms
Atom Decomposition Made Formal min α α 0 s.t. x=𝐃α m n 𝐃 x α min α α 0 s.t. 𝐃α−y 2 ≤ε Approximation Algorithms L0 – counting number of non-zeros in the vector This is a projection onto the Sparseland model These problems are known to be NP-Hard problem Relaxation methods Basis-Pursuit Greedy methods Thresholding/OMP Michael Elad The Computer-Science Department The Technion
21
Pursuit Algorithms min α α 0 s.t. 𝐃α−y 2 ≤ε Basis Pursuit Thresholding
Approximation Algorithms Basis Pursuit Change the L0 into L1 and then the problem becomes convex and manageable Thresholding Multiply y by 𝐃 𝐓 and apply shrinkage: α = 𝒫 𝛽 𝐃 𝐓 y min α α 1 s.t. 𝐃α−y 2 ≤ε Michael Elad The Computer-Science Department The Technion
22
Difficulties with Sparseland
There are various pursuit algorithms Here is an example using the Basis Pursuit (L1): Surprising fact: Many of these algorithms are often accompanied by theoretical guarantees for their success, if the unknown is sparse enough α1 α2 α3 Σ Michael Elad The Computer-Science Department The Technion
23
Assume normalized columns
The Mutual Coherence Compute = 𝐃 𝐃 T Assume normalized columns 𝐃 T 𝐃 The Mutual Coherence μ 𝐃 is the largest off-diagonal entry in absolute value We will pose all the theoretical results in this talk using this property, due to its simplicity You may have heard of other ways to characterize the dictionary (Restricted Isometry Property - RIP, Exact Recovery Condition - ERC, Babel function, Spark, …) Michael Elad The Computer-Science Department The Technion
24
M Basis-Pursuit Success 𝐃
Theorem: Given a noisy signal y=𝐃α+v where v 2 ≤ε and α is sufficiently sparse, then Basis-Pursuit: min α α s.t. 𝐃α−y 2 ≤ ε leads to a stable result: α −α 2 2 ≤ 4 𝜀 2 1−μ 4 α 0 −1 α 0 < μ Donoho, Elad & Temlyakov (‘06) Comments: If =0 α =α This is a worst-case analysis – better bounds exist Similar theorems exist for many other pursuit algorithms x + y v 2 ≤ε 𝐃 min α α 0 s.t. 𝐃α−y 2 ≤ ε min α α 1 s.t. 𝐃α−y 2 ≤ ε α M α Michael Elad The Computer-Science Department The Technion
25
Difficulties with Sparseland
Problem 2: Given a family of signals, how do we find the dictionary to represent it well? Solution: Learn! Gather a large set of signals (many thousands), and find the dictionary that sparsifies them Such algorithms were developed in the past 10 years (e.g., K-SVD), and their performance is surprisingly good We will not discuss this matter further in this talk due to lack of time α1 α2 α3 Σ Michael Elad The Computer-Science Department The Technion
26
Difficulties with Sparseland
Problem 3: Why is this model suitable to describe various sources? e.g., Is it good for images? Audio? Stocks? … General answer: Yes, this model is extremely effective in representing various sources Theoretical answer: Clear connection to other models Empirical answer: In a large variety of signal and image processing (and later machine learning), this model has been shown to lead to state-of-the-art results α1 α2 α3 Σ Michael Elad The Computer-Science Department The Technion
27
ALL ANSWERED POSITIVELY AND CONSTRUCTIVELY
Difficulties with Sparseland ? α1 α2 α3 Σ Problem 1: Given an image patch, how can we find its atom decomposition ? Problem 2: Given a family of signals, how do we find the dictionary to represent it well? Problem 3: Is this model flexible enough to describe various sources? E.g., Is it good for images? audio? … ALL ANSWERED POSITIVELY AND CONSTRUCTIVELY Michael Elad The Computer-Science Department The Technion
28
Sparseland for Image Processing
When handling images, Sparseland is typically deployed on small overlapping patches due to the desire to train the model to fit the data better The model assumption is: each patch in the image is believed to have a sparse representation w.r.t. a common local dictionary What is the corresponding global model? This brings us to … the Convolutional Sparse Coding (CSC) Michael Elad The Computer-Science Department The Technion
29
Multi-Layered Convolutional Sparse Modeling
Joint work with Vardan Papyan Yaniv Romano Jeremias Sulam V. Papyan, J. Sulam, and M. Elad, Working Locally Thinking Globally: Theoretical Guarantees for Convolutional Sparse Coding, IEEE Trans. on Signal Processing, Vol. 65, No. 21, Pages , November 2017. Michael Elad The Computer-Science Department The Technion
30
Convolutional Sparse Coding (CSC)
𝑚 filters convolved with their sparse representations i-th feature-map: An image of the same size as 𝐗 holding the sparse representation related to the i-filter [𝐗]= i=1 𝑚 d i ∗ [Γ i ] An image with 𝑁 pixels The i-th filter of small size 𝑛 This model emerged in , developed and advocated by Yan LeCun and others. It serves as the foundation of Convolutional Neural Networks
31
CSC in Matrix Form Here is an alternative global sparsity-based model formulation 𝐂 i ∈ ℝ 𝑁×𝑁 is a banded and Circulant matrix containing a single atom with all of its shifts 𝚪 i ∈ ℝ 𝑁 are the corresponding coefficients ordered as column vectors 𝐗= i=1 𝑚 𝐂 i 𝚪 i = 𝐂 1 ⋯ 𝐂 𝑚 𝚪 1 ⋮ 𝚪 𝑚 =𝐃𝚪 𝑛 𝑁 𝐂 i =
32
The CSC Dictionary 𝐂 1 𝐂 2 𝐂 3 = 𝑛 𝐃 L 𝑚 𝐃=
33
Why CSC? = 𝐗=𝐃𝚪 𝐑 i 𝐗 𝐑 i+1 𝐗 𝛄 i 𝛄 i+1 Ω stripe vector
𝑛 (2𝑛−1)𝑚 𝑛 (2𝑛−1)𝑚 𝐑 i 𝐗 𝐑 i+1 𝐗 𝛄 i 𝛄 i+1 Ω stripe vector stripe-dictionary 𝐗=𝐃𝚪 Every patch has a sparse representation w.r.t. to the same local dictionary (𝛀) just as assumed for images 𝐑 i 𝐗=𝛀 𝛄 i 𝐑 i+1 𝐗=𝛀 𝛄 i+1
34
Classical Sparse Theory for CSC ?
min 𝚪 𝚪 s.t. 𝐘−𝐃𝚪 2 ≤ε Theorem: BP is guaranteed to “succeed” …. if 𝚪 𝟎 < 𝟏 𝟒 𝟏+ 𝟏 𝛍 Assuming that 𝑚=2 and 𝑛=64 we have that [Welch, ’74] μ≥0.063 Success of pursuits is guaranteed as long as 𝚪 0 < μ(𝐃) ≤ ≈4.2 Only few (4) non-zeros GLOBALLY are allowed!!! This is a very pessimistic result! The classic Sparseland Theory does not provide good explanations for the CSC model
35
Moving to Local Sparsity: Stripes
𝑚=2 ℓ 0,∞ Norm: 𝚪 0,∞ s = max i 𝛄 i 0 The main question we aim to address is this: Can we generalize the vast theory of Sparseland to this new notion of local sparsity? For example, could we provide guarantees for success for pursuit algorithms? min 𝚪 𝚪 0,∞ s s.t. 𝐘−𝐃𝚪 2 ≤ε 𝚪 0,∞ s is low all 𝛄 i are sparse every patch has a sparse representation over 𝛀 𝛄 i 𝛄 i+1 𝚪
36
Success of the Basis Pursuit
Γ BP = min Γ Y−𝐃Γ λ Γ 1 Theorem: For Y=𝐃Γ+E, if λ=4 E 2,∞ p , if 𝚪 𝟎,∞ 𝐬 < 𝟏 𝟑 𝟏+ 𝟏 𝛍 𝐃 then Basis Pursuit performs very-well: The support of Γ BP is contained in that of Γ Γ BP −Γ ∞ ≤7.5 E 2,∞ p Every entry greater than 7.5 E 2,∞ p is found Γ BP is unique This is a much better result – it allows few non-zeros locally in each stripe, implying a permitted O 𝑁 non-zeros globally Papyan, Sulam & Elad (‘17)
37
Multi-Layered Convolutional Sparse Modeling
V. Papyan, Y. Romano, and M. Elad, Convolutional Neural Networks Analyzed via Convolutional Sparse Coding, Journal of Machine Learning Research, Vol. 18, Pages 1-52, July 2017. V. Papyan, Y. Romano, J. Sulam, and M. Elad, Theoretical Foundations of Deep Learning via Sparse Representations, IEEE Signal Processing Magazine, Vol. 35, No. 4, Pages 72-89, June 2018. Vardan Papyan Yaniv Romano Jeremias Sulam Michael Elad The Computer-Science Department The Technion
38
From CSC to Multi-Layered CSC
𝐗∈ ℝ 𝑁 𝑚 1 𝑛 0 𝐃 1 ∈ ℝ 𝑁×𝑁 𝑚 1 𝚪 1 ∈ ℝ 𝑁 𝑚 1 We propose to impose the same structure on the representations themselves 𝚪 1 ∈ ℝ 𝑁 𝑚 1 𝑛 1 𝑚 1 𝑚 2 𝐃 2 ∈ ℝ 𝑁 𝑚 1 ×𝑁 𝑚 2 𝑚 1 𝚪 2 ∈ ℝ 𝑁 𝑚 2 Convolutional sparsity (CSC) assumes an inherent structure is present in natural signals Multi-Layer CSC (ML-CSC)
39
Intuition: From Atoms to Molecules
𝐗 𝐃 1 𝚪 1 𝐃 2 𝚪 2 𝐗 𝐃 1 & The atoms of 𝐃 1 𝐃 2 are combinations of atoms from 𝐃 these are now molecules Thus, this model offers different levels of abstraction in describing X atoms molecules cells tissue body-parts …
40
Intuition: From Atoms to Molecules
𝐃 2 𝐗 𝐃 1 𝐃 K 𝚪 K 𝐃 eff = 𝐃 1 𝐃 2 𝐃 3 ∙∙∙ 𝐃 K 𝐱=𝐃 eff 𝚪 K This is a special Sparseland (indeed, a CSC) model However: A key property in our model: the intermediate representations are required to be sparse as well
41
A Small Taste: Model Training (MNIST)
MNIST Dictionary: D1: 32 filters of size 7×7, with stride of 2 (dense) D2: 128 filters of size 5×5×32 with stride of % sparse D3: 1024 filters of size 7×7×128 – % sparse 𝐃 1 (7×7) 𝐃 1 𝐃 2 (15×15) 𝐃 1 𝐃 2 𝐃 3 (28×28)
42
ML-CSC: Pursuit Deep–Coding Problem 𝐃𝐂 𝐏 λ (dictionaries are known):
Find 𝚪 j j=1 K 𝑠.𝑡. 𝐗= 𝐃 1 𝚪 𝚪 1 0,∞ s ≤ λ 1 𝚪 1 = 𝐃 2 𝚪 𝚪 2 0,∞ s ≤ λ 2 ⋮ ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 K 0,∞ s ≤ λ K Or, more realistically for noisy signals, Find 𝚪 j j=1 K 𝑠.𝑡 𝐘− 𝐃 1 𝚪 ≤ℰ 𝚪 1 0,∞ s ≤ λ 1 𝚪 1 = 𝐃 2 𝚪 𝚪 2 0,∞ s ≤ λ 2 ⋮ ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 K 0,∞ s ≤ λ K
43
A Small Taste: Pursuit x= 𝐃 1 Γ 1 x= 𝐃 1 𝐃 2 Γ 2 x= 𝐃 1 𝐃 2 𝐃 3 Γ 3 x
Y x= 𝐃 1 Γ 1 x= 𝐃 1 𝐃 2 Γ 2 x= 𝐃 1 𝐃 2 𝐃 3 Γ 3 x Γ 0 Γ 1 94.51 % sparse (213 nnz) Γ 2 99.52% sparse (30 nnz) Γ 3 99.51% sparse (5 nnz)
44
ML-CSC: The Simplest Pursuit
The simplest pursuit algorithm (single-layer case) is the THR algorithm, which operates on a given input signal 𝐘 by: 𝐘=𝐃𝚪+𝐄 and 𝚪 is sparse 𝚪 = 𝒫 𝛽 𝐃 T 𝐘
45
Consider this for Solving the DCP
Layered Thresholding (LT): Now let’s take a look at how Conv. Neural Network operates: 𝐃𝐂 𝐏 λ ℰ : Find 𝚪 j j=1 K 𝑠.𝑡. 𝐘− 𝐃 1 𝚪 ≤ℰ 𝚪 1 0,∞ s ≤ λ 1 𝚪 1 = 𝐃 2 𝚪 𝚪 2 0,∞ s ≤ λ 2 ⋮ ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 K 0,∞ s ≤ λ K Estimate 𝚪 1 via the THR algorithm 𝚪 2 = 𝒫 β 2 𝐃 2 T 𝒫 β 1 𝐃 1 T 𝐘 𝚪 2 = 𝒫 β 2 𝐃 2 T 𝒫 β 1 𝐃 1 T 𝐘 𝚪 2 = 𝒫 β 2 𝐃 2 T 𝒫 β 1 𝐃 1 T 𝐘 Estimate 𝚪 2 via the THR algorithm 𝑓 𝐘 =ReLU 𝐛 2 + 𝐖 2 T ReLU 𝐛 1 + 𝐖 1 T 𝐘 The layered (soft nonnegative) thresholding and the CNN forward pass algorithm are the very same thing !!!
46
M A Theoretical Path 𝐗 𝐘 𝚪 i i=1 K 𝐗= 𝐃 1 𝚪 1 𝐃𝐂 𝐏 λ ℰ 𝚪 1 = 𝐃 2 𝚪 2
𝐗= 𝐃 1 𝚪 1 𝚪 1 = 𝐃 2 𝚪 2 ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 i is 𝐋 0,∞ sparse 𝐃𝐂 𝐏 λ ℰ Layered THR (Forward Pass) Maybe other? 𝐗 𝐘 Armed with this view of a generative source model, we may ask new and daring theoretical questions
47
Success of the Layered-THR
Theorem: If 𝚪 i 0,∞ s < μ 𝐃 i ⋅ 𝚪 i min 𝚪 i max − 1 μ 𝐃 i ⋅ ε L i−1 𝚪 i max then the Layered Hard THR (with the proper thresholds) finds the correct supports and 𝚪 i LT − 𝚪 i 2,∞ p ≤ ε L i , where we have defined ε L 0 = 𝐄 2,∞ p and ε L i = 𝚪 i 0,∞ p ⋅ ε L i−1 +μ 𝐃 i 𝚪 i 0,∞ s −1 𝚪 i max Papyan, Romano & Elad (‘17) Problems: Contrast Error growth Error even if no noise The stability of the forward pass is guaranteed if the underlying representations are locally sparse and the noise is locally bounded
48
Layered Basis Pursuit (BP)
𝐃𝐂 𝐏 λ ℰ : Find 𝚪 j j=1 K 𝑠.𝑡. 𝐘− 𝐃 1 𝚪 ≤ℰ 𝚪 1 0,∞ s ≤ λ 1 𝚪 1 = 𝐃 2 𝚪 𝚪 2 0,∞ s ≤ λ 2 ⋮ ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 K 0,∞ s ≤ λ K We chose the Thresholding algorithm due to its simplicity, but we do know that there are better pursuit methods – how about using them? Lets use the Basis Pursuit instead … 𝚪 1 LBP = min 𝚪 𝐘− 𝐃 1 𝚪 λ 1 𝚪 1 1 𝚪 2 LBP = min 𝚪 𝚪 1 LBP − 𝐃 2 𝚪 λ 2 𝚪 2 1 Deconvolutional networks [Zeiler, Krishnan, Taylor & Fergus ‘10]
49
Success of the Layered BP
Theorem: Assuming that 𝚪 𝐢 𝟎,∞ 𝐬 < 𝟏 𝟑 𝟏+ 𝟏 𝛍 𝐃 𝐢 then the Layered Basis Pursuit performs very well: The support of 𝚪 i LBP is contained in that of 𝚪 i The error is bounded: 𝚪 i LBP − 𝚪 i 2,∞ p ≤ ε L i , where ε L i = 7.5 i 𝐄 2,∞ p j=1 i 𝚪 j 0,∞ p Every entry in 𝚪 i greater than ε L i / 𝚪 i 0,∞ p will be found Problems: Contrast Error growth Error even if no noise Papyan, Romano & Elad (‘17)
50
Layered Iterative Thresholding
Layered BP: 𝚪 j LBP = min 𝚪 j 𝚪 j−1 LBP − 𝐃 j 𝚪 j ξ j 𝚪 j 1 j Layered Iterative Soft-Thresholding Algorithm (ISTA): 𝚪 j t = 𝒮 ξ j / c j 𝚪 j t−1 + 𝐃 j T 𝚪 j−1 − 𝐃 j 𝚪 j t−1 t j Note that our suggestion implies that groups of layers share the same dictionaries Can be seen as a very deep residual neural network [He, Zhang, Ren, & Sun ‘15]
51
M Where are the Labels? 𝐗 Answer 1: 𝐗= 𝐃 1 𝚪 1 𝚪 1 = 𝐃 2 𝚪 2 ⋮
We do not need labels because everything we show refer to the unsupervised case, in which we operate on signals, not necessarily in the context of recognition M 𝐗 𝐗= 𝐃 1 𝚪 1 𝚪 1 = 𝐃 2 𝚪 2 ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 i is 𝐋 0,∞ sparse We presented the ML-CSC as a machine that produces signals X
52
M Where are the Labels? 𝐗 Answer 2: 𝐗= 𝐃 1 𝚪 1 𝚪 1 = 𝐃 2 𝚪 2 ⋮
In fact, this model could be augmented by a synthesis of the corresponding label by: L 𝐗 =𝑠𝑖𝑔𝑛 c+ j=1 K w j T Γ j This assumes that knowing the representations suffices for classification supervised mode Thus, a successful pursuit algorithm can lead to an accurate recognition if the network is augmented by a FC classification layer In fact, we can analyze theoretically the classification accuracy and the sensitivity to adversarial noise – see later M 𝐗 𝐗= 𝐃 1 𝚪 1 𝚪 1 = 𝐃 2 𝚪 2 ⋮ 𝚪 K−1 = 𝐃 K 𝚪 K 𝚪 i is 𝐋 0,∞ sparse L 𝐗 We presented the ML-CSC as a machine that produces signals X 55
53
What About Learning? Sparseland Sparse Representation Theory CSC
Convolutional Sparse Coding ML-CSC Multi-Layered Convolutional Sparse Coding All these models rely on proper Dictionary Learning Algorithms to fulfil their mission: Sparseland: We have unsupervised and supervised such algorithms, and a beginning of theory to explain how these work CSC: We have few and only unsupervised methods, and even these are not fully stable/clear ML-CSC: Two algorithms were proposed – unsupervised and supervised 56
54
Time to Conclude
55
Sparseland The desire to model data This Talk Take Home Message 1: Generative modeling of data sources enables algorithm development along with theoretically analyzing algorithms’ performance Novel View of Convolutional Sparse Coding Take Home Message 2: The Multi-Layer Convolutional Sparse Coding model could be a new platform for understanding and developing deep-learning solutions Multi-Layer Convolutional Sparse Coding A novel interpretation and theoretical understanding of CNN The ML-CSC was shown to enable a theoretical study of CNN, along with new insights We spoke about the importance of models in signal/image processing and described Sparseland in details We presented a theoretical study of the CSC model and how to operate locally while getting global optimality We propose a multi-layer extension of CSC, shown to be tightly connected to CNN
56
On a Personal Note … Disclaimer: True, I am biased, so take what I say with a grain of salt … My research activity in the foreseen future will be dedicated to establishing this connection, addressing various aspects of it Conjecture: Sparse modeling of data is at the heart of the Deep-Learning ideas, and as such, it is the main avenue for developing theoretical foundations for this field. Elad (‘19) Michael Elad The Computer-Science Department The Technion
57
Fresh from the Oven My team’s work proceeds along the above-described line of thinking: J. Sulam, V. Papyan, Y. Romano, and M. Elad, Multi-Layer Convolutional Sparse Modeling: Pursuit and Dictionary Learning, IEEE Trans. on Signal Proc., Vol. 66, No. 15, Pages , August 2018. A. Aberdam, J. Sulam, and M. Elad, Multi Layer Sparse Coding: the Holistic Way, SIAM Journal on Mathematics of Data Science (SIMODS), Vol. 1, No. 1, Pages J. Sulam, A. Aberdam, A. Beck, and M. Elad, On Multi-Layer Basis Pursuit, Efficient Algorithms and Convolutional Neural Networks, to appear in IEEE T-PAMI. Y. Romano, A. Aberdam, J. Sulam, and M. Elad, Adversarial Noise Attacks of Deep Learning Architectures – Stability Analysis via Sparse Modeled Signals, to appear in JMIV. Ev Zisselman, Jeremias Sulam, and Michael Elad, A Local Block Coordinate Descent Algorithm for the CSC Model, CVPR 2019. I. Rey-Otero, J. Sulam, and M. Elad, Variations on the CSC model, submitted to IEEE Transactions on Signal Processing. D. Simon and M. Elad, Rethinking the CSC model for Natural Images, submitted to NIPS. 64
58
A New Massive Open Online Course
Michael Elad The Computer-Science Department The Technion
59
Questions? More on these (including these slides and the relevant papers) can be found in
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.