Algorithm for non-negative matrix factorization Daniel D. Lee, H. Sebastian Seung. Algorithm for non-negative matrix factorization. Nature.

Slides:



Advertisements
Similar presentations
Nonnegative Matrix Factorization with Sparseness Constraints S. Race MA591R.
Advertisements

Numerical Solution of Linear Equations
Scientific Computing QR Factorization Part 2 – Algorithm to Find Eigenvalues.
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Separating Hyperplanes
Solving Linear Systems (Numerical Recipes, Chap 2)
Section 1.7 Diagonal, Triangular, and Symmetric Matrices.
Autar Kaw Benjamin Rigsby Transforming Numerical Methods Education for STEM Undergraduates.
SOLVING SYSTEMS OF LINEAR EQUATIONS. Overview A matrix consists of a rectangular array of elements represented by a single symbol (example: [A]). An individual.
Rayan Alsemmeri Amseena Mansoor. LINEAR SYSTEMS Jacobi method is used to solve linear systems of the form Ax=b, where A is the square and invertible.
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
Linear Transformations
Numerical Optimization
Motion Analysis (contd.) Slides are from RPI Registration Class.
Methods For Nonlinear Least-Square Problems
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
Matrix Operations. Matrix Notation Example Equality of Matrices.
Learning the parts of objects by nonnegative matrix factorization D.D. Lee from Bell Lab H.S. Seung from MIT Presenter: Zhipeng Zhao.
Unconstrained Optimization Problem
6 1 Linear Transformations. 6 2 Hopfield Network Questions.
Linear Least Squares QR Factorization. Systems of linear equations Problem to solve: M x = b Given M x = b : Is there a solution? Is the solution unique?
Why Function Optimization ?
Today Wrap up of probability Vectors, Matrices. Calculus
9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
By Mary Hudachek-Buswell. Overview Atmospheric Turbulence Blur.
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
Scientific Computing QR Factorization Part 1 – Orthogonal Matrices, Reflections and Rotations.
Collaborative Filtering Matrix Factorization Approach
Length and Dot Product in R n Notes: is called a unit vector. Notes: The length of a vector is also called its norm. Chapter 5 Inner Product Spaces.
Chapter 8 Objectives Understanding matrix notation.
SINGLE CHANNEL SPEECH MUSIC SEPARATION USING NONNEGATIVE MATRIXFACTORIZATION AND SPECTRAL MASKS Jain-De,Lee Emad M. GraisHakan Erdogan 17 th International.
Scientific Computing Linear Systems – LU Factorization.
8.1 Vector spaces A set of vector is said to form a linear vector space V Chapter 8 Matrices and vector spaces.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
Dr. Mubashir Alam King Saud University. Outline Systems of Linear Equations (6.1) Matrix Arithmetic (6.2) Arithmetic Operations (6.2.1) Elementary Row.
1 Iterative Solution Methods Starts with an initial approximation for the solution vector (x 0 ) At each iteration updates the x vector by using the sytem.
Non Negative Matrix Factorization
Linear algebra: matrix Eigen-value Problems
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
CHAPTER 2 MATRICES 2.1 Operations with Matrices Matrix
Lecture 8 Matrix Inverse and LU Decomposition
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 4. Least squares.
Adaptive Algorithms for PCA PART – II. Oja’s rule is the basic learning rule for PCA and extracts the first principal component Deflation procedure can.
Chap. 5 Inner Product Spaces 5.1 Length and Dot Product in R n 5.2 Inner Product Spaces 5.3 Orthonormal Bases: Gram-Schmidt Process 5.4 Mathematical Models.
A Clustering Method Based on Nonnegative Matrix Factorization for Text Mining Farial Shahnaz.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Lecture 10 - Nonlinear gradient techniques and LU Decomposition CVEN 302 June 24, 2002.
NONNEGATIVE MATRIX FACTORIZATION WITH MATRIX EXPONENTIATION Siwei Lyu ICASSP 2010 Presenter : 張庭豪.
Chapter 2-OPTIMIZATION
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Krylov-Subspace Methods - I Lecture 6 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Gaoal of Chapter 2 To develop direct or iterative methods to solve linear systems Useful Words upper/lower triangular; back/forward substitution; coefficient;
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Maximum Norms & Nonnegative Matrices  Weighted maximum norm e.g.) x1x1 x2x2.
MA5233 Lecture 6 Krylov Subspaces and Conjugate Gradients Wayne M. Lawton Department of Mathematics National University of Singapore 2 Science Drive 2.
Krylov-Subspace Methods - II Lecture 7 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Feature Generation and Cluster-based Feature Selection.
Matrices. Variety of engineering problems lead to the need to solve systems of linear equations matrixcolumn vectors.
Linear Algebra by Dr. Shorouk Ossama.
Krylov-Subspace Methods - I
Linear Algebra Lecture 15.
Mikhail Bilenko, Sugato Basu, Raymond J. Mooney
RECORD. RECORD Gaussian Elimination: derived system back-substitution.
Collaborative Filtering Matrix Factorization Approach
~ Least Squares example
~ Least Squares example
Performance Optimization
Lecture 8 Matrix Inverse and LU Decomposition
Non-Negative Matrix Factorization
Presentation transcript:

Algorithm for non-negative matrix factorization Daniel D. Lee, H. Sebastian Seung. Algorithm for non-negative matrix factorization. Nature.

Unsupervised learning algorithms Unsupervised learning algorithms: –PCA –VQ (vector quantization) –NMF

NMF –Triangular Factorization (LU) cannot work. It should be a square matrix. –QR Factorization MATLAB use qr function to do QR Factorization, the syntax is –[Q,R]= qr(A) –Q is a orthonormal matrix QQ T =I –R is a upper triangle matrix –A is a m*n matrix –iterative m*rr*nr < n or m

Cost function A,B are two non-negative matrices Count the distance between A and B –Euclidean distance: –Kullback-Leibler distance (Relative Entropy)

Cost function These two equations are both lower bounded by zero. Proof

Goal Minimize with respect to and subject to the constraints, Gradient descent : –simplest technique to implement. –convergence can be slow. –inconvenient for large applications. Conjugate gradient : –Faster convergence, at least in the vicinity of local minima. –Complicated to implement.

Multiplicative update rules Euclidean distance Divergence

Multiplicative versus additive update rules

Definition 1: is an auxiliary function for F(h) if the conditions are satisfied. Lemma 1: proof: Note: only if is a local minimum of

Lemma 2 If is the diagonal matrix

Proof