A Single-letter Characterization of Optimal Noisy Compressed Sensing Dongning Guo Dror Baron Shlomo Shamai
Setting Replace samples by more general measurements based on a few linear projections (inner products) measurements sparse signal # non-zeros
Signal Model Signal entry X n = B n U n iid B n » Bernoulli() sparse iid U n » P U PUPU Bernoulli() Multiplier PXPX
Measurement Noise Measurement process is typically analog Analog systems add noise, non-linearities, etc. Assume Gaussian noise for ease of analysis Can be generalized to non-Gaussian noise
Noiseless measurements denoted y 0 Noise Noisy measurements Unit-norm columns SNR= Noise Model noiseless SNR
Model process as measurement channel Measurements provide information! channel CS measurement CS decoding source encoder channel encoder channel decoder source decoder Allerton 2006 [Sarvotham, Baron, & Baraniuk]
Theorem: [Sarvotham, Baron, & Baraniuk 2006] For sparse signal with rate-distortion function R(D), lower bound on measurement rate s.t. SNR and distortion D Numerous single-letter bounds –[Aeron, Zhao, & Saligrama] –[Akcakaya and Tarokh] –[Rangan, Fletcher, & Goyal] –[Gastpar & Reeves] –[Wang, Wainwright, & Ramchandran] –[Tune, Bhaskaran, & Hanly] –… Single-Letter Bounds
Goal: Precise Single-letter Characterization of Optimal CS
What Single-letter Characterization? Ultimately what can one say about X n given Y? (sufficient statistic) Very complicated Want a simple characterization of its quality Large-system limit: channelposterior
Main Result: Single-letter Characterization Result1: Conditioned on X n =x n, the observations (Y,) are statistically equivalent to easy to compute… Estimation quality from (Y,) just as good as noisier scalar observation degradation channelposterior
2 (0,1) is fixed point of Take-home point: degraded scalar channel Non-rigorous owing to replica method w/ symmetry assumption –used in CDMA detection [Tanaka 2002, Guo & Verdu 2005] Related analysis [Rangan, Fletcher, & Goyal 2009] –MMSE estimate (not posterior) using [Guo & Verdu 2005] –extended to several CS algorithms particularly LASSO Details
Decoupling
Result2: Large system limit; any arbitrary (constant) L input elements decouple: Take-home point: “interference” from each individual signal entry vanishes Decoupling Result
Sparse Measurement Matrices
Sparse Measurement Matrices [Baron, Sarvotham, & Baraniuk] LDPC measurement matrix (sparse) Mostly zeros in ; nonzeros » P Each row contains ¼ Nq randomly placed nonzeros Fast matrix-vector multiplication fast encoding / decoding sparse matrix
CS Decoding Using BP [Baron, Sarvotham, & Baraniuk] Measurement matrix represented by graph Estimate input iteratively Implemented via nonparametric BP [Bickson,Sommer,…] measurements y signal x
Identical Single-letter Characterization w/BP Result3: Conditioned on X n =x n, the observations (Y,) are statistically equivalent to Sparse matrices just as good BP is asymptotically optimal! identical degradation
Decoupling Between Two Input Entries (N=500, M=250, =0.1, =10) density
CS-BP vs Other CS Methods (N=1000, =0.1, q=0.02) MM MMSE CS-BP
Conclusion Single-letter characterization of CS Decoupling Sparse matrices just as good Asymptotically optimal CS-BP algorithm