Recovering structured signals: Precise performance analysis Christos Thrampoulidis Joint ITA Workshop, La Jolla, CA February 3, 2016.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Regularized risk minimization
Joint work with Irad Yavneh
Pixel Recovery via Minimization in the Wavelet Domain Ivan W. Selesnick, Richard Van Slyke, and Onur G. Guleryuz *: Polytechnic University, Brooklyn, NY.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
List decoding Reed-Muller codes up to minimal distance: Structure and pseudo- randomness in coding theory Abhishek Bhowmick (UT Austin) Shachar Lovett.
The Stability of a Good Clustering Marina Meila University of Washington
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
CMPUT 466/551 Principal Source: CMU
Approximate Message Passing for Bilinear Models
ECE Department Rice University dsp.rice.edu/cs Measurements and Bits: Compressed Sensing meets Information Theory Shriram Sarvotham Dror Baron Richard.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Sparse and Overcomplete Data Representation
Richard Baraniuk Rice University dsp.rice.edu/cs Compressive Signal Processing.
Compressed Sensing for Networked Information Processing Reza Malek-Madani, 311/ Computational Analysis Don Wagner, 311/ Resource Optimization Tristan Nguyen,
Efficient Statistical Pruning for Maximum Likelihood Decoding Radhika Gowaikar Babak Hassibi California Institute of Technology July 3, 2003.
A Single-letter Characterization of Optimal Noisy Compressed Sensing Dongning Guo Dror Baron Shlomo Shamai.
Compressive Signal Processing
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Location Estimation in Sensor Networks Moshe Mishali.
Nonlinear Sampling. 2 Saturation in CCD sensors Dynamic range correction Optical devices High power amplifiers s(-t) Memoryless nonlinear distortion t=n.
Lattices for Distributed Source Coding - Reconstruction of a Linear function of Jointly Gaussian Sources -D. Krithivasan and S. Sandeep Pradhan - University.
1 Chapter 1 Introduction. 2 Outline 1.1 A Very Abstract Summary 1.2 History 1.3 Model of the Signaling System 1.4 Information Source 1.5 Encoding a Source.
Representation and Compression of Multi-Dimensional Piecewise Functions Dror Baron Signal Processing and Systems (SP&S) Seminar June 2009 Joint work with:
Compression with Side Information using Turbo Codes Anne Aaron and Bernd Girod Information Systems Laboratory Stanford University Data Compression Conference.
Hybrid Dense/Sparse Matrices in Compressed Sensing Reconstruction
Topics in MMSE Estimation for Sparse Approximation Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa 32000,
Cs: compressed sensing
Physics 270 – Experimental Physics. Standard Deviation of the Mean (Standard Error) When we report the average value of n measurements, the uncertainty.
1 Sparsity Control for Robust Principal Component Analysis Gonzalo Mateos and Georgios B. Giannakis ECE Department, University of Minnesota Acknowledgments:
Learning With Structured Sparsity
Abhik Majumdar, Rohit Puri, Kannan Ramchandran, and Jim Chou /24 1 Distributed Video Coding and Its Application Presented by Lei Sun.
Channel Capacity.
Extremal Problems of Information Combining Alexei Ashikhmin  Information Combining: formulation of the problem  Mutual Information Function for the Single.
Toward a Secure Data-Rate Theorem Paul Cuff. Control Setting Controller Encoder System (Plant) Sensors Rate R UiUi XiXi YiYi.
Compressible priors for high-dimensional statistics Volkan Cevher LIONS/Laboratory for Information and Inference Systems
Constrained adaptive sensing Mark A. Davenport Georgia Institute of Technology School of Electrical and Computer Engineering TexPoint fonts used in EMF.
Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Instructor: Yonina Eldar Teaching Assistant: Tomer Michaeli Spring 2009 Modern Sampling Methods
BCS547 Neural Decoding.
High-dimensional Error Analysis of Regularized M-Estimators Ehsan AbbasiChristos ThrampoulidisBabak Hassibi Allerton Conference Wednesday September 30,
Robust Principal Components Analysis IT530 Lecture Notes.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Zhilin Zhang, Bhaskar D. Rao University of California, San Diego March 28,
Cameron Rowe.  Introduction  Purpose  Implementation  Simple Example Problem  Extended Kalman Filters  Conclusion  Real World Examples.
A Low-Complexity Universal Architecture for Distributed Rate-Constrained Nonparametric Statistical Learning in Sensor Networks Avon Loy Fernandes, Maxim.
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
1 On the Channel Capacity of Wireless Fading Channels C. D. Charalambous and S. Z. Denic School of Information Technology and Engineering, University of.
Robust Optimization and Applications in Machine Learning.
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
Regularized Least-Squares and Convex Optimization.
Super-resolution MRI Using Finite Rate of Innovation Curves Greg Ongie*, Mathews Jacob Computational Biomedical Imaging Group (CBIG) University of Iowa.
Today.
Ch3: Model Building through Regression
Regularized risk minimization
Lecture 09: Gaussian Processes
Matrix Completion from a few entries
Distributed Submodular Maximization in Massive Datasets
Probabilistic Models for Linear Regression
USPACOR: Universal Sparsity-Controlling Outlier Rejection
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Sudocodes Fast measurement and reconstruction of sparse signals
Slides prepared on 09 September, 2018
Biointelligence Laboratory, Seoul National University
Lecture 10: Gaussian Processes
Sudocodes Fast measurement and reconstruction of sparse signals
Subspace Expanders and Low Rank Matrix Recovery
Presentation transcript:

Recovering structured signals: Precise performance analysis Christos Thrampoulidis Joint ITA Workshop, La Jolla, CA February 3, 2016

Let’s start “simple”…

Given y and A can you find x 0 ? Noiseless linear measurements

Does it work? (if so, when?) Let A be random (e.g. Gaussian, random DFT). [Candes,Romberg,Tao’04] then, x 0 is the unique solution with high probability. If,

Suppose A has entries iid Gaussian. Please, be precise… Then, [Donoho’06; Donoho,Tanner’09; Stojnic’09; Donoho et.al.’09; Amelunxen et.al.’13; Stojnic’13] In fact, much more is known… Generalization to other structures (e.g. group-sparse, low-rank, binary entries, smooth) and convex regularizer functions f (e.g. l 1,2, nuclear-norm, atomic-norms). Gaussian-width, statistical dimension [Donoho et. al.’09; Chandrasekaran et.al.’12; Amelunxen et.al.’13]

Crystal clear understanding!

One step further…

Add some noise!

Can you estimate x 0 ? Noisy linear measurements

What is the quality of the estimate? LASSO Suppose A has randomness. then w.h.p.: and If e.g. [Belloni’11]

Precise results?

Precise results? Yes! [Donoho, Maleki, Montanari’09] and [Bayati, Montanari’11]. Approach is based on the AMP framework, inspired by statistical physics. Over the past couple of year, we have developed a new framework the CGMT that is based on Gaussian process methods. Framework is inspired by [Stojnic’13]. Gives results that are very general. Natural way of analysis… CGMT

Precise results? Yes! [C.T., Panahi, Guo, Hassibi’14]

Precise results? Yes! [C.T., Panahi, Guo, Hassibi’14]

Precise results? Yes! [C.T., Panahi, Guo, Hassibi’14]

Precise results? Yes! [C.T., Panahi, Guo, Hassibi’14]

Precise results? Yes! [C.T., Panahi, Guo, Hassibi’14]

Precise results? Yes! Expected Moreau Envelope summary functional: captures the role of regularizer and of p X0. generalizes the “Gaussian-width” of noiseless CS. [Thrampoulidis, Panahi, Guo, Hassibi’14]

Optimal tuning There is value to be precise! Noiseless CS results follow as special case:

Generalizations General structures and convex regularizers. Other performance metrics (e.g. support recovery, BER) – Massive MIMO (BER of convex relaxation decoders) General noise distributions and loss functions! Regularized M-estimators

Master Theorem [Thrampoulidis, Abbasi, Hassibi ’16]

Optimal tuning Precise is good!

Optimal loss function? 23 Precise & general is better!

Consistent Estimators? 24 sparse noise

Robustness to outliers? Huber loss function + Cauchy noise

Final challenge

measurement device has nonlinearities & uncertainties. can also arise by design Non-linear measurements (or single-index model)

measurement device has nonlinearities & uncertainties. can also arise by design Non-linear measurements (or single-index model)

What if we use the LASSO? 1.the link function might be unknown/misspecified (robustness) 2.widely available LASSO solvers, used in practice 3.turns out it performs well

Theorem (Non-linear=linear) For example, for 1-bit measurements: [Thrampoulidis, Abbasi, Hassibi. NIPS’15] Previous work of Plan & Vershynin ‘15 derives order-wise bounds for the constrained LASSO. Our precise result explicitly captures equivalence & the role of σ.

Application: optimal quantization q-bit measurements LASSO Lloyd-Max (LM) algorithm is optimal A LASSO A

One last word: How it all works?

Gaussian min-max Theorem (Gordon’88) Convex Gaussian Min-max Theorem [Thrampoulidis, Oymak, Hassibi. COLT ’15] CGMT framework

Gaussian min-max Theorem (Gordon’88) Convex Gaussian Min-max Theorem [Thrampoulidis, Oymak, Hassibi. COLT ’15] CGMT framework GMT

Thrampoulidis, Abbasi, Hassibi: “Precise Error Analysis of Regularized M-estimators in High-dimensions”, available on arXiv, Jan More… Thank you!