Commentary on “The Characterization, Subtraction, and Addition of Astronomical Images” by Robert Lupton Rebecca Willett.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Active Appearance Models
PARAMETER ESTIMATION FOR ODES USING A CROSS-ENTROPY APPROACH Wayne Enright Bo Wang University of Toronto.
Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Component Analysis (Review)
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Data Mining Feature Selection. Data reduction: Obtain a reduced representation of the data set that is much smaller in volume but yet produces the same.
Eigen-analysis and the Power Method
PCA + SVD.
Multiscale Analysis of Photon-Limited Astronomical Images Rebecca Willett.
Image-Pro Premier Basic Training Course
, Tim Landgraf Active Appearance Models AG KI, Journal Club 03 Nov 2008.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
L15:Microarray analysis (Classification). The Biological Problem Two conditions that need to be differentiated, (Have different treatments). EX: ALL (Acute.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Ordinary least squares regression (OLS)
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Dan Simon Cleveland State University
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Summarized by Soo-Jin Kim
Principle Component Analysis (PCA) Networks (§ 5.8) PCA: a statistical procedure –Reduce dimensionality of input vectors Too many features, some of them.
Probability of Error Feature vectors typically have dimensions greater than 50. Classification accuracy depends upon the dimensionality and the amount.
Presented By Wanchen Lu 2/25/2013
Cao et al. ICML 2010 Presented by Danushka Bollegala.
Minimum Phoneme Error Based Heteroscedastic Linear Discriminant Analysis for Speech Recognition Bing Zhang and Spyros Matsoukas BBN Technologies Present.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Dimension reduction : PCA and Clustering Slides by Agnieszka Juncker and Chris Workman modified by Hanne Jarmer.
Image deconvolution, denoising and compression T.E. Gureyev and Ya.I.Nesterets
Dimensionality Reduction Motivation I: Data Compression Machine Learning.
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
Math 5364/66 Notes Principal Components and Factor Analysis in SAS Jesse Crawford Department of Mathematics Tarleton State University.
Cluster Analysis Potyó László. Cluster: a collection of data objects Similar to one another within the same cluster Similar to one another within the.
Separating Style and Content with Bilinear Models Joshua B. Tenenbaum, William T. Freeman Computer Examples Barun Singh 25 Feb, 2002.
A Semi-Blind Technique for MIMO Channel Matrix Estimation Aditya Jagannatham and Bhaskar D. Rao The proposed algorithm performs well compared to its training.
Discriminant Analysis
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 10: PRINCIPAL COMPONENTS ANALYSIS Objectives:
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –(Finish) Expectation Maximization –Principal Component Analysis (PCA) Readings:
Irena Váňová. B A1A1. A2A2. A3A3. repeat until no sample is misclassified … labels of classes Perceptron algorithm for i=1...N if then end * * * * *
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Hybrid Bayesian Linearized Acoustic Inversion Methodology PhD in Petroleum Engineering Fernando Bordignon Introduction Seismic inversion.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
CSE 554 Lecture 8: Alignment
Compressive Coded Aperture Video Reconstruction
LECTURE 11: Advanced Discriminant Analysis
Background on Classification
Principle Component Analysis (PCA) Networks (§ 5.8)
School of Computer Science & Engineering
LECTURE 10: DISCRIMINANT ANALYSIS
Blind Signal Separation using Principal Components Analysis
Course Outline MODEL INFORMATION COMPLETE INCOMPLETE
Probabilistic Models with Latent Variables
Lecture 14 PCA, pPCA, ICA.
'Linear Hierarchical Models'
PCA is “an orthogonal linear transformation that transfers the data to a new coordinate system such that the greatest variance by any projection of the.
Parallelization of Sparse Coding & Dictionary Learning
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Feature space tansformation methods
Separating Style and Content with Bilinear Models Joshua B
LECTURE 09: DISCRIMINANT ANALYSIS
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Separating Style and Content with Bilinear Models Joshua B
Presentation transcript:

Commentary on “The Characterization, Subtraction, and Addition of Astronomical Images” by Robert Lupton Rebecca Willett

Focus of commentary KL transform and data scarcity Improved PSF estimation via blind deconvolution

First principal component Second principal component Principal Components Analysis (aka KL Transform) 1.Compute sample covariance matrix (pXp) 2.Determine directions of greatest variance using eigenanalysis

Principal Components Analysis (aka KL Transform) Key advantages: 1.Optimal linear method for dimensionality reduction 2.Model parameters computed directly from data 3.Reduction and expansion easy to compute

Data Scarcity When using the KL to estimate the PSF, –p (dimension of data) = 120 –n (number of point sources observed) = 20 p >> n What effect does this have when performing PCA? –Sample covariance matrix not full rank –Need special care in implementation –Naïve computational complexity O(np 2 )

Working around the data scarcity problem Preprocess data by performing dimensionality reduction (Johnstone & Lu, 2004) Use an EM algorithm to solve for k-term PSF; O(knp) complexity (Roweis 1998) Balance between decorrelation and sparsity (Chennubholta & Jepson, 2001) PCA Sparse PCA

Blind Deconvolution Advantages Not necessary to pick out “training” stars Potential to use prior knowledge of image structure/statistics Possible to estimate distended PSF features (e.g. ghosting effects) Potential to use information from multiple exposures Disadvantages Computational complexity can be prohibitive Can be overkill if only PSF, and not deconvolved image, is desired

Example of blind deconvolution: modified Richardson-Lucy 1.Start with initial intensity image estimate and initial PSF estimate 2.R-L update of intensity given PSF 3.R-L update of PSF estimate given intensity 4.Goto 2 (depends on good initial estimates) Tsumuraya, Miura, & Baba 1993

Iterative error minimization Minimize this function: Jefferies & Christou, 1993

Simulation example Iterative Blind Deconvolution Weiner Deconvolution Maximum Entropy Deconvolution Observations

Data from multiple exposures H1H1 y1y1 = Poisson H2H2 y2y2 H3H3 y3y3 If H i = H. S i, where H is the imager PSF and S i is a known shift operator, then we can use multiple exposures to more accurately estimate H.

Takeaway messages Exercise caution when using the KL transform to estimate the PSF –Avoid computing sample covariance matrix –Consider iterative, low computational complexity methods Blind deconvolution indirectly estimates PSF –Uses prior knowledge of image structure/statistics –Requires less arbitrary user input –Can estimate non-local PSF components