Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.

Slides:



Advertisements
Similar presentations
Ordinary Least-Squares
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
Matrix Factorization with Unknown Noise
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Data Modeling and Parameter Estimation Nov 9, 2005 PSCI 702.
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
Ch11 Curve Fitting Dr. Deshi Ye
Chapter 6 Feature-based alignment Advanced Computer Vision.
The General Linear Model. The Simple Linear Model Linear Regression.
1 Chapter 4 Interpolation and Approximation Lagrange Interpolation The basic interpolation problem can be posed in one of two ways: The basic interpolation.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Coefficient Path Algorithms Karl Sjöstrand Informatics and Mathematical Modelling, DTU.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Curve-Fitting Regression
Epipolar geometry. (i)Correspondence geometry: Given an image point x in the first view, how does this constrain the position of the corresponding point.
Structure from motion. Multiple-view geometry questions Scene geometry (structure): Given 2D point matches in two or more images, where are the corresponding.
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Algorithm Evaluation and Error Analysis class 7 Multiple View Geometry Comp Marc Pollefeys.
Kalman Filtering Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
Bootstrapping a Heteroscedastic Regression Model with Application to 3D Rigid Motion Evaluation Bogdan Matei Peter Meer Electrical and Computer Engineering.
CSCE 641 Computer Graphics: Image-based Modeling (Cont.) Jinxiang Chai.
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
Matrix Approach to Simple Linear Regression KNNL – Chapter 5.
Collaborative Filtering Matrix Factorization Approach
Computer vision: models, learning and inference
Chapter 6 Feature-based alignment Advanced Computer Vision.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
CSCE 643 Computer Vision: Structure from Motion
Curve-Fitting Regression
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
Direct Robust Matrix Factorization Liang Xiong, Xi Chen, Jeff Schneider Presented by xxx School of Computer Science Carnegie Mellon University.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
Introduction to Matrices and Matrix Approach to Simple Linear Regression.
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Efficient Gaussian Process Regression for large Data Sets ANJISHNU BANERJEE, DAVID DUNSON, SURYA TOKDAR Biometrika, 2008.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
An Introduction To The Kalman Filter By, Santhosh Kumar.
1.3 Solutions of Linear Systems
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
University of Colorado Boulder ASEN 5070 Statistical Orbit determination I Fall 2012 Professor George H. Born Professor Jeffrey S. Parker Lecture 9: Least.
EECS 274 Computer Vision Projective Structure from Motion.
Fitting.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Numerical Analysis – Data Fitting Hanyang University Jong-Il Park.
École Doctorale des Sciences de l'Environnement d’Île-de-France Année Universitaire Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
Reconstruction of a Scene with Multiple Linearly Moving Objects Mei Han and Takeo Kanade CISC 849.
Lecture 16: Image alignment
12. Principles of Parameter Estimation
Answering ‘Where am I?’ by Nonlinear Least Squares
Motion Segmentation with Missing Data using PowerFactorization & GPCA
CH 5: Multivariate Methods
Structure from motion Input: Output: (Tomasi and Kanade)
Matrices Definition: A matrix is a rectangular array of numbers or symbolic elements In many applications, the rows of a matrix will represent individuals.
Collaborative Filtering Matrix Factorization Approach
Regression Models - Introduction
12. Principles of Parameter Estimation
Structure from motion Input: Output: (Tomasi and Kanade)
Regression Models - Introduction
Presentation transcript:

Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel CVPR 2010

Usual low rank approximation using L 2 norm– SVD. Robust low rank approximation using L 2 norm- Wiberg Algorithm. “Robust” low rank approximation in the presence of: – missing data –Outliers –L 1 norm –Generalization of Wiberg Algorithm. Y= UV MXN MXR RXN

Problem W is the indicator matrix, w ij = 1 if y ij is known, else 0.

Wiberg Algorithm W matrix indicates the presence/absence of elements From: “On the Wiberg algorithm for matrix factorization in the presence of missing components”, Okatani et al, IJCV 2006,

Alternating Least Squares To find the minimum of φ, find derivatives Considering the two equations independently. Starting with some initial estimates u 0 and v 0, update u from v and v from u. Converges very slowly, specially for missing components and strong noise. From: “On the Wiberg algorithm for matrix factorization in the presence of missing components”, Okatani et al, IJCV 2006,

Back to Wiberg In non-linear least squares problems with multiple parameters, when assuming part of the parameters to be fixed, minimization of the least squares with respect to the rest of the parameters becomes a simple problem, e.g., a linear problem. So closed form solutions may be found. Wiberg applied it to this problem of factorization of matrix with missing components. From: “On the Wiberg algorithm for matrix factorization in the presence of missing components”, Okatani et al, IJCV 2006,

Back to Wiberg For a fixed u, the L 2 norm becomes a linear, least squares minimization problem in v. –Compute optimal v*(u) Apply Gauss-Newton method to the above non-linear least squares problem to find optimal u*. Easy to compute derivative because of L 2 norm From: “On the Wiberg algorithm for matrix factorization in the presence of missing components”, Okatani et al, IJCV 2006,

Linear Programming and Definitions

L 1 -Wiberg Algorithm Minimization problem in terms of L 1 norm Minimization problem in terms of v and u independently Substituting v* into u

Comparing to L 2 -Wiberg V*(U) is not easily differentiable The minimization function (u,v*) is not a least squares minimization problem, so Gauss-Newton can’t be applied directly. Idea: Let V*(U) denote the optimal basic solution. V*(U) is differentiable assuming problem is feasible, as per Fundamental Theorem of differentiability of linear programs. Jacobian for the G-N :: derivative of solution to a linear prog. problem

≈ Add an additional term to the function and minimize the value of the term ?

Results Tested on synthetic data. –Randomly created measurement matrices Y drawn from a uniform distribution [-1,1]. –20% missing, 10% noise [-5,5]. Real data –Dinosaur sequence from oxford-vgg.

Structure from motion Projections of 319 points tracked over 36 views. Addition of noise to 10% points. Full 3d reconstruction ~ low rank matrix approximation. Above-residual for the visible points. In L 2 norm, reconstruction error is evenly distributed among all elements of residual. In L 1 norm, error concentrated on few elements.