E XACT MATRIX C OMPLETION VIA CONVEX OPTIMIZATION E MMANUEL J. C ANDES AND B ENJAMIN R ECHT M AY 2008 Presenter: Shujie Hou January, 28 th,2011 Department.

Slides:



Advertisements
Similar presentations
Totally Unimodular Matrices
Advertisements

The Stability of a Good Clustering Marina Meila University of Washington
T HE POWER OF C ONVEX R ELAXATION : N EAR - OPTIMAL MATRIX COMPLETION E MMANUEL J. C ANDES AND T ERENCE T AO M ARCH, 2009 Presenter: Shujie Hou February,
Extremum Properties of Orthogonal Quotients Matrices By Achiya Dax Hydrological Service, Jerusalem, Israel
by Rianto Adhy Sasongko Supervisor: Dr.J.C.Allwright
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Lecture 19 Singular Value Decomposition
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Sampling algorithms for l 2 regression and applications Michael W. Mahoney Yahoo Research (Joint work with P. Drineas.
1 Latent Semantic Indexing Jieping Ye Department of Computer Science & Engineering Arizona State University
Support Vector Classification (Linearly Separable Case, Primal) The hyperplanethat solves the minimization problem: realizes the maximal margin hyperplane.
Classification Problem 2-Category Linearly Separable Case A- A+ Malignant Benign.
Singular Value Decomposition
Matrix sparsification and the sparse null space problem Lee-Ad GottliebWeizmann Institute Tyler NeylonBynomial Inc. TexPoint fonts used in EMF. Read the.
Clustering In Large Graphs And Matrices Petros Drineas, Alan Frieze, Ravi Kannan, Santosh Vempala, V. Vinay Presented by Eric Anderson.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Lecture 20 SVD and Its Applications Shang-Hua Teng.
Singular Value Decomposition and Data Management
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
On L1q Regularized Regression Authors: Han Liu and Jian Zhang Presented by Jun Liu.
Probability theory 2008 Outline of lecture 5 The multivariate normal distribution  Characterizing properties of the univariate normal distribution  Different.
1cs542g-term Notes  Extra class next week (Oct 12, not this Friday)  To submit your assignment: me the URL of a page containing (links to)
T he Separability Problem and its Variants in Quantum Entanglement Theory Nathaniel Johnston Institute for Quantum Computing University of Waterloo.
Low-Rank Solution of Convex Relaxation for Optimal Power Flow Problem
Introduction The central problems of Linear Algebra are to study the properties of matrices and to investigate the solutions of systems of linear equations.
Fundamentals from Linear Algebra Ghan S. Bhatt and Ali Sekmen Mathematical Sciences and Computer Science College of Engineering Tennessee State University.
+ Review of Linear Algebra Optimization 1/14/10 Recitation Sivaraman Balakrishnan.
Matrix Completion IT530 Lecture Notes.
Cs: compressed sensing
© 2005 Yusuf Akgul Gebze Institute of Technology Department of Computer Engineering Computer Vision Geometric Camera Calibration.
C&O 355 Mathematical Programming Fall 2010 Lecture 4 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Chapter 3 – Set Theory  .
Orthogonality and Least Squares
Orthogonalization via Deflation By Achiya Dax Hydrological Service Jerusalem, Israel
CPSC 491 Xin Liu November 19, Norm Norm is a generalization of Euclidean length in vector space Definition 2.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Section 2.3 Properties of Solution Sets
A Note on Rectangular Quotients By Achiya Dax Hydrological Service Jerusalem, Israel
C&O 355 Mathematical Programming Fall 2010 Lecture 16 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Review of Linear Algebra Optimization 1/16/08 Recitation Joseph Bradley.
Robust Principal Components Analysis IT530 Lecture Notes.
Charles University FSV UK STAKAN III Institute of Economic Studies Faculty of Social Sciences Institute of Economic Studies Faculty of Social Sciences.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
1 Chapter 7 Numerical Methods for the Solution of Systems of Equations.
Optimal Reverse Prediction: Linli Xu, Martha White and Dale Schuurmans ICML 2009, Best Overall Paper Honorable Mention A Unified Perspective on Supervised,
Krylov-Subspace Methods - I Lecture 6 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
Matrices, Vectors, Determinants.
Part 3 Linear Programming 3.3 Theoretical Analysis.
Dimension reduction (2) EDR space Sliced inverse regression Multi-dimensional LDA Partial Least Squares Network Component analysis.
Commuting birth-and-death processes Caroline Uhler Department of Statistics UC Berkeley (joint work with Steven N. Evans and Bernd Sturmfels) MSRI Workshop.
The Duality Theorem Primal P: Maximize
Non-additive Security Games
Sam Hopkins Cornell Tselil Schramm UC Berkeley Jonathan Shi Cornell
Matrix Completion from a few entries
LSI, SVD and Data Management
Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Presenter: Xia Li.
Chap 9. General LP problems: Duality and Infeasibility
Nuclear Norm Heuristic for Rank Minimization
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Singular Value Decomposition SVD
Rank-Sparsity Incoherence for Matrix Decomposition
Maths for Signals and Systems Linear Algebra in Engineering Lecture 6, Friday 21st October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
Back to Cone Motivation: From the proof of Affine Minkowski, we can see that if we know generators of a polyhedral cone, they can be used to describe.
CIS 700: “algorithms for Big Data”
(Convex) Cones Def: closed under nonnegative linear combinations, i.e.
Presentation transcript:

E XACT MATRIX C OMPLETION VIA CONVEX OPTIMIZATION E MMANUEL J. C ANDES AND B ENJAMIN R ECHT M AY 2008 Presenter: Shujie Hou January, 28 th,2011 Department of Electrical and Computer Engineering Cognitive Radio Institute Tennessee Technological university

S OME A VAILABLE C ODES rank/sample_code.html#RPCA e.html

O UTLINE The problem statement Examples of impossible recovery Algorithms Main theorems Proof Experimental results Discussion

PROBLEM C ONSIDERED The problem of low-rank matrix completion: Recovery of a data matrix from a sampling of its entries. A matrix with rows and columns Only observing a number of of its entries which is much smaller than. Can the partially observed matrix be recovered and under what kind of conditions such a matrix can be exactly recovered?

E XAMPLES OF IMPOSSIBLE RECOVERY This matrix can not be recovered unless all of the entries are given. Reason: for most sampling sets, only observing all of zeros. Not all of the matrices can be completed from a sample of their entries.

E XAMPLES OF IMPOSSIBLE R ECOVERY The observation does not include samples from first row, the first component could never be guessed out. Not all the sampling set can be used to complete the matrix.

S HORT CONCLUSION One can not recover all low-rank matrices from any set of sampled entries. Can one recover most matrices from almost all sampling sets of cardinality ? The two theorems given later will tell that this is possible for most low-rank matrices under some specific conditions.

A LGORITHM Intuitively, (NP-hard problem) Alternatively, considering a heuristic optimization in which Is (1.5) reasonable or to what extent it is equivalent to rank minimization? Locations of observed entries Sum of singular values

T HE FIRST T HEOREM There is unique low-rank matrix consistent with the observed entries. The heuristic model (1.5) is equivalent to the above NP-hard formulation. Talk later

R ANDOM O RTHOGONAL M ODEL Generic low-rank matrix SVD (singular value decomposition ) of a matrix

A DEFINITION The subspace with low coherence is the special interest of this paper. Singular vectors with low coherence is “ spread out.”(not sparse) It can guarantee that the sampling set cannot really be a zero set.

T WO ASSUMPTIONS

M AIN RESULTS Theorem1.1 is a special case of theorem 1.3. If only a few matrices satisfy the conditions, it will also make the theorem 1.3 of little practical use.

T HE CONDITIONS OF THEOREM 1.3 The random orthogonal model obeys the two assumptions A0 and A1 with large probability.

T HE PROOF OF THE LEMMAS

R EVIEW OF T HEOREM 1.3

T HE PROOF The author employs the tools of subgradient (in distributions (generalized functional) and duality ( in optimization theory) and tools in asymptotic geometric analysis to prove the existence and uniqueness of the theorem (1.3). The proof is from page The details won’t be discussed here.

D UALITY (1) In optimization theory, the duality principle states that optimization problems may be viewed from either of two perspectives, the primal problem or the dual problem. (Wikipedia) Concept in constraint optimization

D UALITY (2) The largest singular value

D UALITY (3) in which Orthogonal complement

D UALITY (4) Ensure Y is the subgradient of nuclear norm at X0 Equivalent to the operator Property of injectivity

A RCHITECTURE OF THE PROOF (1) The candidate Y which vanishes on the complement of the will be the solution to the optimization model of

A RCHITECTURE OF THE PROOF (2) The candidate Y which vanishes on the complement of the will be the solution to The first part of the statement 1. Hopefully, the small Frobenius norm will indicate the small spectral norm as well. Prove the first statement

A RCHITECTURE OF THE PROOF (3) Ready to prove the second property: injectivity of Property of the orthogonal projection If is a one-to-one linear mapping, then The solution of the model:

A RCHITECTURE OF THE PROOF (4) Bernoulli model of the sampling set

A RCHITECTURE OF THE PROOF (5)

A RCHITECTURE OF THE P ROOF (6)

C ONNECTIONS TO TRACE HEURISTIC When the matrix variable is symmetric and positive semidefinite: Which is equivalent to

E XTENSIONS The matrix completion can be extended to multitask and multiclass learning problems in machine learning.

N UMERICAL EXPERIMENTS

N UMERICAL RESULTS

D ISCUSSIONS Under suitable conditions, the matrix can be completed for a small number of the sampled entries. The required number of the sample entries is on the order of.

Questions? Thank you!