NONNEGATIVE MATRIX FACTORIZATION WITH MATRIX EXPONENTIATION Siwei Lyu ICASSP 2010 Presenter : 張庭豪.

Slides:



Advertisements
Similar presentations
Nonnegative Matrix Factorization with Sparseness Constraints S. Race MA591R.
Advertisements

Accelerated, Parallel and PROXimal coordinate descent IPAM February 2014 APPROX Peter Richtárik (Joint work with Olivier Fercoq - arXiv: )
Edge Preserving Image Restoration using L1 norm
Modulation Spectrum Factorization for Robust Speech Recognition Wen-Yi Chu 1, Jeih-weih Hung 2 and Berlin Chen 1 Presenter : 張庭豪.
1 Transportation problem The transportation problem seeks the determination of a minimum cost transportation plan for a single commodity from a number.
P. Venkataraman Mechanical Engineering P. Venkataraman Rochester Institute of Technology DETC2013 – 12269: Continuous Solution for Boundary Value Problems.
Dragan Jovicic Harvinder Singh
SVM—Support Vector Machines
Structured Sparse Principal Component Analysis Reading Group Presenter: Peng Zhang Cognitive Radio Institute Friday, October 01, 2010 Authors: Rodolphe.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
P. Venkataraman Mechanical Engineering P. Venkataraman Rochester Institute of Technology DETC2014 – 35148: Continuous Solution for Boundary Value Problems.
Nonsmooth Nonnegative Matrix Factorization (nsNMF) Alberto Pascual-Montano, Member, IEEE, J.M. Carazo, Senior Member, IEEE, Kieko Kochi, Dietrich Lehmann,
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
by Rianto Adhy Sasongko Supervisor: Dr.J.C.Allwright
More MR Fingerprinting
An Introduction to Sparse Coding, Sparse Sensing, and Optimization Speaker: Wei-Lun Chao Date: Nov. 23, 2011 DISP Lab, Graduate Institute of Communication.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Coefficient Path Algorithms Karl Sjöstrand Informatics and Mathematical Modelling, DTU.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Martin Burger Institut für Numerische und Angewandte Mathematik European Institute for Molecular Imaging CeNoS Total Variation and related Methods Numerical.
1 Chapter 8: Linearization Methods for Constrained Problems Book Review Presented by Kartik Pandit July 23, 2010 ENGINEERING OPTIMIZATION Methods and Applications.
NORM BASED APPROACHES FOR AUTOMATIC TUNING OF MODEL BASED PREDICTIVE CONTROL Pastora Vega, Mario Francisco, Eladio Sanz University of Salamanca – Spain.
Unconstrained Optimization Problem
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
Lecture 10: Support Vector Machines
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
1 Multiple Kernel Learning Naouel Baili MRL Seminar, Fall 2009.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Informatics and Mathematical Modelling / Intelligent Signal Processing ISCAS Morten Mørup Approximate L0 constrained NMF/NTF Morten Mørup Informatics.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.

IE 607 Constrained Design: Using Constraints to Advantage in Adaptive Optimization in Manufacturing.
Chapter 3 Linear Programming Methods 高等作業研究 高等作業研究 ( 一 ) Chapter 3 Linear Programming Methods (II)
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Column Generation Approach for Operating Rooms Planning Mehdi LAMIRI, Xiaolan XIE and ZHANG Shuguang Industrial Engineering and Computer Sciences Division.
Optimization for Operation of Power Systems with Performance Guarantee
Non Negative Matrix Factorization
Cs: compressed sensing
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Barnett/Ziegler/Byleen Finite Mathematics 11e1 Learning Objectives for Section 6.4 The student will be able to set up and solve linear programming problems.
Local Non-Negative Matrix Factorization as a Visual Representation Tao Feng, Stan Z. Li, Heung-Yeung Shum, HongJiang Zhang 2002 IEEE Presenter : 張庭豪.
Machine Learning Using Support Vector Machines (Paper Review) Presented to: Prof. Dr. Mohamed Batouche Prepared By: Asma B. Al-Saleh Amani A. Al-Ajlan.
To accompany Quantitative Analysis for Management, 8e by Render/Stair/Hanna 11-1 © 2003 by Prentice Hall, Inc. Upper Saddle River, NJ Chapter 11.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Nonlinear Programming Models
Distributed Nonnegative Matrix Factorization for Web- Scale Dyadic Data Analysis on MapReduce Challenge : the scalability of available tools Definition.
Low-Rank Kernel Learning with Bregman Matrix Divergences Brian Kulis, Matyas A. Sustik and Inderjit S. Dhillon Journal of Machine Learning Research 10.
Linear Programming Erasmus Mobility Program (24Apr2012) Pollack Mihály Engineering Faculty (PMMK) University of Pécs João Miranda
Chapter 6 Linear Programming: The Simplex Method Section 4 Maximization and Minimization with Problem Constraints.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Monte-Carlo method for Two-Stage SLP Lecture 5 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO Working Group on Continuous.
Learning Spectral Clustering, With Application to Speech Separation F. R. Bach and M. I. Jordan, JMLR 2006.
A Clustering Method Based on Nonnegative Matrix Factorization for Text Mining Farial Shahnaz.
Endmember Extraction from Highly Mixed Data Using MVC-NMF Lidan Miao AICIP Group Meeting Apr. 6, 2006 Lidan Miao AICIP Group Meeting Apr. 6, 2006.
A Convergent Solution to Tensor Subspace Learning.
Ariadna Quattoni Xavier Carreras An Efficient Projection for l 1,∞ Regularization Michael Collins Trevor Darrell MIT CSAIL.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Instructional Design Document Simplex Method - Optimization STAM Interactive Solutions.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Non-negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs C.G. Puntonet and A. Prieto (Eds.): ICA 2004 Presenter.
Regularized Least-Squares and Convex Optimization.
StingyCD: Safely Avoiding Wasteful Updates in Coordinate Descent
Multiplicative updates for L1-regularized regression
C.-S. Shieh, EC, KUAS, Taiwan
Lecture 8 – Nonlinear Programming Models
1.3 Vector Equations.
Non-Negative Matrix Factorization
Outline Sparse Reconstruction RIP Condition
Presentation transcript:

NONNEGATIVE MATRIX FACTORIZATION WITH MATRIX EXPONENTIATION Siwei Lyu ICASSP 2010 Presenter : 張庭豪

Outline ABSTRACT NMF AND CONSTRAINED NMF NMF WITH MATRIX EXPONENTIATION EXPERIMENTS CONCLUSION 2

ABSTRACT  Nonnegative matrix factorization (NMF) has been successfully applied to different domains as a technique able to find part-based linear representations for nonnegative data.  However, when extra constraints are incorporated into NMF, simple gradient descent optimization can be inefficient for high dimensional problems, due to the overhead to enforce the nonnegativity constraints. 3

ABSTRACT  We describe an alternative formulation based on matrix exponentiation, where the nonnegativity constraints are enforced implicitly, and a direct gradient descent algorithm can have better efficiency.  In numerical experiments, such a reformulation leads to significant improvement in running time. 4

NMF AND CONSTRAINED NMF  Nonnegative matrix factorization (NMF) aims to find a concise representation for nonnegative data.  Given an m × n data matrix V with V i j ≥ 0 and an integer k ≤ min( m, n ), NMF finds two nonnegative factors W ∈ R m × k and H ∈ R k × n so that V ≈ WH.  The nonnegativity constraints make the representation purely additive that affords a part-based interpretation. 5

NMF AND CONSTRAINED NMF  Mathematically, NMF is formulated as minimizing a loss function L(W, H; V), subject to the nonnegativity constraints.  Two most commonly used objective functions :  (1) squared Euclidean distance :  (2) matrix divergence 6

NMF AND CONSTRAINED NMF  The original NMF algorithm used multiplicative update steps, as  We loosely refer to such variations of NMF that have extra constraints on the nonnegative factors as constrained NMF.  For instance, in problems such as face recognition, a part-based representation is also preferably a sparse representation, where a small set of factors with non-zero weights suffice to reconstruct an objective.  This leads to the development of sparse NMF. 7

NMF AND CONSTRAINED NMF  A general way to introduce these further requirements on the NMF factors is to add penalty terms in the NMF objective function.  Nevertheless, when these new objective functions are used, the multiplicative update steps are no longer viable,as they do not guarantee to decrease the modified objective function of the constrained NMF.  The procedure guarantees to converge to a local minimum of the objective function, yet its efficiency in high-dimensional problem is greatly affected by these clipping operations. 8

NMF WITH MATRIX EXPONENTIATION  We reformulate NMF to find factors that are the point-wise exponentiation of another matrices.  The introduction of matrix exponentiation annihilates the nonnegativity constraints,and a gradient descent with matrix exponentiation always searches the optimal solution within the feasible set, thus avoids moving on the constraint boundaries.  Furthermore, gradient descent with matrix exponentiation can also make greater progress in regions relatively faraway from the optimal solution. 9

NMF WITH MATRIX EXPONENTIATION 10

NMF WITH MATRIX EXPONENTIATION  The derivatives of L2(W, H; V) with regards to W and H are computed as: 11

NMF WITH MATRIX EXPONENTIATION  In terms of numerical optimization techniques, matrix exponentiation formulation of NMF can be viewed as an interior-point style solution to the bounded optimization problem in the NMF, with the exponentiation serves as the barrier function.  The matrix exponentiation formulation can also be employed in other constrained NMF algorithms where the original multiplicative updates are not applicable. 12

EXPERIMENTS Compare :  Gradient descent with matrix exponentiation (ME)  Original NMF update algorithm (NMF)  Projected gradient descent of the (constrained) NMF objective function without any transformation (PGD)  We tested the algorithm on both artificial data, where V is a randomly generated 100 × 100 positive matrix, and we search for factors with k = 25, and a real data set from the a subset of the USPS handwritten digit database. 13

EXPERIMENTS Sparse NMF  In the first set of experiments, we enforce the sparseness requirement to the NMF problem by adding penalty term in the NMF objective function that penalizes nonsparse decompositions.  A particular effective measurement of sparseness of an n-dimensional vector x is given in, as which is 0 for a vector with equal entries ( most nonsparse ) and 1 for a vector with only one nonzero entry ( most sparse ). 14

EXPERIMENTS 15 ( μ W and μ H are small positive constants (stepsizes) ) (L1) (L2) fix the L2 norm of h i to unity

EXPERIMENTS 16 s i = x i – (∑x i – L1)/dim(x)

EXPERIMENTS 17 s m 45 度 s-m 圓與直線去求一 元二次方程式求 ( 大於 0 的 )

EXPERIMENTS  We used an augmented objective function, which penalizes nonsparse columns in W and H T, with λ W and λ H constants controlling the balance between data fitting and sparsity requirements in the factors. 18

EXPERIMENTS Smooth NMF  Smooth NMF is a constrained variant of NMF where one seeks nonnegative factors that are also smooth, i.e., having small difference between consecutive elements in the factors.  Smooth NMF has application in analyzing fMRI data,where a desirable property of the recovered factors is their smoothness in the temporal and voxel domain. 19

EXPERIMENTS  objective function: where D is a difference operator. Less smooth columns in W and H T are penalized by the of l 2 norm of their differentials. 20

CONCLUSION  In this paper, we describe a new formulation of NMF and its constrained variants with matrix exponentiation.  The major advantage of this new formulation is that nonnegativity constraints are enforced implicitly and thus obviate the overhead of a gradient descent algorithm to check the nonnegativity constraints.  Furthermore, the nonlinear transformation of the objective function introduced by matrix exponentiationcan also accelerate convergence speed in regions that are faraway from the optimal solution, which also helps to reduce the overall running time of constrained NMF. 21

NMF WITH MATRIX EXPONENTIATION  The derivatives of D(W, H; V) with regards to W and H are computed as: 22