Multi-Task Compressive Sensing with Dirichlet Process Priors Yuting Qi 1, Dehong Liu 1, David Dunson 2, and Lawrence Carin 1 1 Department of Electrical.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Compressive Sensing IT530, Lecture Notes.
Gibbs Sampling Methods for Stick-Breaking priors Hemant Ishwaran and Lancelot F. James 2001 Presented by Yuting Qi ECE Dept., Duke Univ. 03/03/06.
Pixel Recovery via Minimization in the Wavelet Domain Ivan W. Selesnick, Richard Van Slyke, and Onur G. Guleryuz *: Polytechnic University, Brooklyn, NY.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Computer vision: models, learning and inference Chapter 8 Regression.
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
Chapter 10 Curve Fitting and Regression Analysis
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
Exact or stable image\signal reconstruction from incomplete information Project guide: Dr. Pradeep Sen UNM (Abq) Submitted by: Nitesh Agarwal IIT Roorkee.
Learning sparse representations to restore, classify, and sense images and videos Guillermo Sapiro University of Minnesota Supported by NSF, NGA, NIH,
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
More MR Fingerprinting
An Introduction to Sparse Coding, Sparse Sensing, and Optimization Speaker: Wei-Lun Chao Date: Nov. 23, 2011 DISP Lab, Graduate Institute of Communication.
Compressed sensing Carlos Becker, Guillaume Lemaître & Peter Rennert
Learning With Dynamic Group Sparsity Junzhou Huang Xiaolei Huang Dimitris Metaxas Rutgers University Lehigh University Rutgers University.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
A Bayesian Approach to Joint Feature Selection and Classifier Design Balaji Krishnapuram, Alexander J. Hartemink, Lawrence Carin, Fellow, IEEE, and Mario.
“Random Projections on Smooth Manifolds” -A short summary
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Richard Baraniuk Rice University dsp.rice.edu/cs Compressive Signal Processing.
Compressed Sensing for Networked Information Processing Reza Malek-Madani, 311/ Computational Analysis Don Wagner, 311/ Resource Optimization Tristan Nguyen,
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
1 Unsupervised Learning With Non-ignorable Missing Data Machine Learning Group Talk University of Toronto Monday Oct 4, 2004 Ben Marlin Sam Roweis Rich.
Introduction to Compressive Sensing
Markus Strohmeier Sparse MRI: The Application of
Rice University dsp.rice.edu/cs Distributed Compressive Sensing A Framework for Integrated Sensing and Processing for Signal Ensembles Marco Duarte Shriram.
Computer Science Department Andrés Corrada-Emmanuel and Howard Schultz Presented by Lawrence Carin from Duke University Autonomous precision error in low-
Department of Computer Engineering University of California at Santa Cruz Data Compression (2) Hai Tao.
Empirical Bayes approaches to thresholding Bernard Silverman, University of Bristol (joint work with Iain Johnstone, Stanford) IMS meeting 30 July 2002.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Informatics and Mathematical Modelling / Intelligent Signal Processing 1 EUSIPCO’09 27 August 2009 Tuning Pruning in Sparse Non-negative Matrix Factorization.
The horseshoe estimator for sparse signals CARLOS M. CARVALHO NICHOLAS G. POLSON JAMES G. SCOTT Biometrika (2010) Presented by Eric Wang 10/14/2010.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Hierarchical Dirichelet Processes Y. W. Tech, M. I. Jordan, M. J. Beal & D. M. Blei NIPS 2004 Presented by Yuting Qi ECE Dept., Duke Univ. 08/26/05 Sharing.
Compressive Sensing Based on Local Regional Data in Wireless Sensor Networks Hao Yang, Liusheng Huang, Hongli Xu, Wei Yang 2012 IEEE Wireless Communications.
Cs: compressed sensing
Learning With Structured Sparsity
Learning Lateral Connections between Hidden Units Geoffrey Hinton University of Toronto in collaboration with Kejie Bao University of Toronto.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Shriram Sarvotham Dror Baron Richard Baraniuk ECE Department Rice University dsp.rice.edu/cs Sudocodes Fast measurement and reconstruction of sparse signals.
Estimating Component Availability by Dempster-Shafer Belief Networks Estimating Component Availability by Dempster-Shafer Belief Networks Lan Guo Lane.
MITSUBISHI ELECTRIC RESEARCH LABORATORIES Cambridge, Massachusetts High resolution SAR imaging using random pulse timing Dehong Liu IGARSS’ 2011 Vancouver,
CHAPTER 5 SIGNAL SPACE ANALYSIS
Regularization and Feature Selection in Least-Squares Temporal Difference Learning J. Zico Kolter and Andrew Y. Ng Computer Science Department Stanford.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Lecture 2: Statistical learning primer for biologists
Zhilin Zhang, Bhaskar D. Rao University of California, San Diego March 28,
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Multi-label Prediction via Sparse Infinite CCA Piyush Rai and Hal Daume III NIPS 2009 Presented by Lingbo Li ECE, Duke University July 16th, 2010 Note:
APPLICATIONS OF DIRICHLET PROCESS MIXTURES TO SPEAKER ADAPTATION Amir Harati and Joseph PiconeMarc Sobel Institute for Signal and Information Processing,
Fundamentals of Data Analysis Lecture 10 Correlation and regression.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Sparsity Based Poisson Denoising and Inpainting
Bayesian Semi-Parametric Multiple Shrinkage
Chapter 7. Classification and Prediction
Compressive Coded Aperture Video Reconstruction
Probability Theory and Parameter Estimation I
Highly Undersampled 0-norm Reconstruction
Ch3: Model Building through Regression
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Analyzing Redistribution Matrix with Wavelet
Learning With Dynamic Group Sparsity
Filtering and State Estimation: Basic Concepts
Sudocodes Fast measurement and reconstruction of sparse signals
INFONET Seminar Application Group
Sudocodes Fast measurement and reconstruction of sparse signals
Learning Incoherent Sparse and Low-Rank Patterns from Multiple Tasks
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Presentation transcript:

Multi-Task Compressive Sensing with Dirichlet Process Priors Yuting Qi 1, Dehong Liu 1, David Dunson 2, and Lawrence Carin 1 1 Department of Electrical and Computer Engineering 2 Department of Statistical Science Duke University, USA

Overviews of CS - 1/6 Nyquist Sampling  f sampling >= 2 f max  In many applications, f sampling is very high.  Most digital signals are highly compressible, only encode few large coefficients and throw away most of them. WT 512*512 pixels Wavelet coefficients IWT

Overviews of CS - 2/6  Why waste so many measurements if eventually most are discarded?  A surprising experiment: FT Randomly throw away 83% of samples Shepp-Logan phantom A convex non-linear reconstruction

Overviews of CS - 3/6 Basic Idea ( Donoho, Candes, Tao, et al ):  Assume an compressible signal, with an orthonormal basis and θ sparse coefficients.  In CS, we measure v, a compact form of signal u, T is with elements constituted randomly. measurements Sparse signal N: # non-zeros

Overviews of CS - 4/6  The theory of Candes et al. (2006): With overwhelming probability, θ (hence u ) is recovered with if the number of CS measurements ( C is a constant and N is number of non-zeros in θ ) If N is small, i.e., u is highly compressible, m<<n. The problem may be solved by linear programming or greedy algorithms.

Overviews of CS - 5/6 Bayesian CS ( Ji and Carin, 2007 )  Recall  Connection to linear regression  BCS Put sparse prior over θ, Given observation v, p(θ|v) ?

Overviews of CS - 6/6 Multi-Task CS  M CS tasks.  Reduce measurement number by exploiting relationships among tasks.  Existing methods assume all tasks fully shared.  In practice, not all signals are satisfied with this assumption.  Can we expect an algorithm simultaneously discovers sharing structure of all tasks and perform the CS inversion of the underlying signals within each group?

DP Multi-task CS - 1/4 DP MT CS  M sets of CS measurements  May be from different scenarios. some are heart MRI, some are skeleton MRI.  What we want? Share information among all sets of CS tasks when sharing is appropriate. Reduce measurement number. v i : CS measurements of i-th task θ i : underlying sparse signal of i-th task Φ i : random projection matrix of i-th task є i : measurement error of i-th task

DP Multi-task CS - 2/4 DP MT CS Formula  Put sparseness prior over sparse signal θ i :  Encourage sharing of α i, variance of sparse prior, via DP prior:  If necessary, then sharing; otherwise, no.  Estimate signals and learn sharing structure automatically & simultaneously.

DP Multi-Task CS - 3/4  Choice of G 0 S parseness promoting prior:   Automatic relevance determination (ARD) prior which enforces the sparsity over parameters.  If c=d, this becomes a student-t distribution t(0,1).

DP Multi-Task CS - 4/4 Mathematical representation: Stick-breaking components Hyper priors over parameters Association Sparseness Prior

Inference Variational Bayesian Inference  Bayes rule:  Introduce q(Φ) to approximate p(Φ|X,Ψ) ;  Log marginal likelihood  q(Φ) is obtained by maximizing, which is computationally tractable. Marginal likelihood = ? + min max

Experiments - 1/6 Synthetic data  Data are generated from 10 underlying clusters.  Each cluster is generated from one signal template.  Each template has a length of 256, with 30 spikes drawn from N(0,1); locations are random too.  Correlation of any two templates is zero.  From each template, we generate 5 signals: random move 3 spikes.  Total 50 sparse signals.

Experiments - 2/6 Figure.2 (a) Reconstruction error: DPMT CS and fully sharing MT CS (100 runs). (b) Histogram of number of clusters inferred from DPMT CS (100 runs).

Experiments - 3/6 Figure.3 Five underlying clustersFigure.4 Three underlying clusters Figure.5 Two underlying clustersFigure.6 One underlying cluster

Experiments - 4/6  Interesting observations: As # of underlying clusters decreases, the difference of DP-MT and global-sharing MT CS decreases. Sparseness sharing means sharing non-zero components AND zero components Each cluster has distinct non-zero components, BUT they share large amount of zero components. One global sparseness prior is enough to describe two clusters by treating them as one cluster. However, for ten-cluster case, ten templates do not cumulatively share the same set of zero-amplitude coefficients; So global sparseness prior is inappropriate. Cluster 1 Cluster 2

Experiments - 5/6 Real image  12 images of 256 by 256, Sparse in wavelet domain.  Image reconstruction: Collect CS measurements (random projection of θ); estimate θ via CS inversion; reconstruct image by inverse wavelet transformation.  Hybrid scheme: Assume finest wavelet coefficients zero, only estimate other 4096 coefficients (θ); Assume all coarsest coefficients are measured; CS measurements are performed on coefficients other than finest and coarsest ones. Wavelet coefficients IWT Collect CS measurements CS inversion

Table 1 Reconstruction Error

Conclusions A DP-based multi-task compressive sensing framework is developed for jointly performing multiple CS inversion tasks. This new method can simultaneously discover sharing structure of all tasks and perform the CS inversion of the underlying signals within each group. A variational Bayesian inference is developed for computational efficiency. Both synthetic and real image data show MT CS works at least as well as ST CS and outperforms full-sharing MT CS when the assumption of full-sharing is not true.

Thanks for your attention! Questions?