Online Learning for Matrix Factorization and Sparse Coding

Slides:



Advertisements
Similar presentations
Nonnegative Matrix Factorization with Sparseness Constraints S. Race MA591R.
Advertisements

The economical eigen-computation is where the sufficient spanning set is The final updated LDA component is given by Motivation Incremental Linear Discriminant.
Semi-Stochastic Gradient Descent Peter Richtárik ANC/DTC Seminar, School of Informatics, University of Edinburgh Edinburgh - November 4, 2014.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Online PLCA for Real-Time Semi-supervised Source Separation Zhiyao Duan, Gautham J. Mysore, Paris Smaragdis 1. EECS Department, Northwestern University.
Semi-Stochastic Gradient Descent Methods Jakub Konečný (joint work with Peter Richtárik) University of Edinburgh.
Semi-Stochastic Gradient Descent Methods Jakub Konečný University of Edinburgh BASP Frontiers Workshop January 28, 2014.
Sparse Modeling for Finding Representative Objects Ehsan Elhamifar Guillermo Sapiro Ren´e Vidal Johns Hopkins University University of Minnesota Johns.
Structured Sparse Principal Component Analysis Reading Group Presenter: Peng Zhang Cognitive Radio Institute Friday, October 01, 2010 Authors: Rodolphe.
Nonsmooth Nonnegative Matrix Factorization (nsNMF) Alberto Pascual-Montano, Member, IEEE, J.M. Carazo, Senior Member, IEEE, Kieko Kochi, Dietrich Lehmann,
Image Congealing (batch/multiple) image (alignment/registration) Advanced Topics in Computer Vision (048921) Boris Kimelman.
Patch-based Image Deconvolution via Joint Modeling of Sparse Priors Chao Jia and Brian L. Evans The University of Texas at Austin 12 Sep
Proximal Methods for Sparse Hierarchical Dictionary Learning Rodolphe Jenatton, Julien Mairal, Guillaume Obozinski, Francis Bach Presented by Bo Chen,
Learning sparse representations to restore, classify, and sense images and videos Guillermo Sapiro University of Minnesota Supported by NSF, NGA, NIH,
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Efficient Sparse Coding Algorithms
Reinforcement Learning
* Joint work with Michal Aharon Guillermo Sapiro
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Semi-Stochastic Gradient Descent Methods Jakub Konečný (joint work with Peter Richtárik) University of Edinburgh SIAM Annual Meeting, Chicago July 7, 2014.
Relevance Feedback Content-Based Image Retrieval Using Query Distribution Estimation Based on Maximum Entropy Principle Irwin King and Zhong Jin The Chinese.
Sparse and Redundant Representation Modeling for Image Processing Michael Elad The Computer Science Department The Technion – Israel Institute of technology.
(Fri) Young Ki Baik Computer Vision Lab.
Image Classification using Sparse Coding: Advanced Topics
Online Dictionary Learning for Sparse Coding International Conference on Machine Learning, 2009 Julien Mairal, Francis Bach, Jean Ponce and Guillermo Sapiro.
SINGLE CHANNEL SPEECH MUSIC SEPARATION USING NONNEGATIVE MATRIXFACTORIZATION AND SPECTRAL MASKS Jain-De,Lee Emad M. GraisHakan Erdogan 17 th International.
Olshausen’s Demo. 1.The Training set ?  Natural Images (Olhausen’s database)  How much do we learn ?  face database and car database 2.The Sparseness.
Object Bank Presenter : Liu Changyu Advisor : Prof. Alex Hauptmann Interest : Multimedia Analysis April 4 th, 2013.
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
Non Negative Matrix Factorization
Selective Block Minimization for Faster Convergence of Limited Memory Large-scale Linear Models Kai-Wei Chang and Dan Roth Experiment Settings Block Minimization.
Learning a Fast Emulator of a Binary Decision Process Center for Machine Perception Czech Technical University, Prague ACCV 2007, Tokyo, Japan Jan Šochman.
BrainStorming 樊艳波 Outline Several papers on icml15 & cvpr15 PALM Information Theory Learning.
 Karthik Gurumoorthy  Ajit Rajwade  Arunava Banerjee  Anand Rangarajan Department of CISE University of Florida 1.
Local Non-Negative Matrix Factorization as a Visual Representation Tao Feng, Stan Z. Li, Heung-Yeung Shum, HongJiang Zhang 2002 IEEE Presenter : 張庭豪.
Online Learning for Collaborative Filtering
Group Sparse Coding Samy Bengio, Fernando Pereira, Yoram Singer, Dennis Strelow Google Mountain View, CA (NIPS2009) Presented by Miao Liu July
Center for Evolutionary Functional Genomics Large-Scale Sparse Logistic Regression Jieping Ye Arizona State University Joint work with Jun Liu and Jianhui.
Greedy is not Enough: An Efficient Batch Mode Active Learning Algorithm Chen, Yi-wen( 陳憶文 ) Graduate Institute of Computer Science & Information Engineering.
Intelligent Database Systems Lab Presenter: YAN-SHOU SIE Authors: RAMIN PASHAIE, STUDENT MEMBER, IEEE, AND NABIL H. FARHAT, LIFE FELLOW, IEEE TNN.
Distributed Nonnegative Matrix Factorization for Web- Scale Dyadic Data Analysis on MapReduce Challenge : the scalability of available tools Definition.
Zibin Zheng DR 2 : Dynamic Request Routing for Tolerating Latency Variability in Cloud Applications CLOUD 2013 Jieming Zhu, Zibin.
Learning to Sense Sparse Signals: Simultaneous Sensing Matrix and Sparsifying Dictionary Optimization Julio Martin Duarte-Carvajalino, and Guillermo Sapiro.
Non-local Sparse Models for Image Restoration Julien Mairal, Francis Bach, Jean Ponce, Guillermo Sapiro and Andrew Zisserman ICCV 2009 Presented by: Mingyuan.
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Intelligent Control and Automation, WCICA 2008.
A Clustering Method Based on Nonnegative Matrix Factorization for Text Mining Farial Shahnaz.
An Efficient Greedy Method for Unsupervised Feature Selection
Design of PCA and SVM based face recognition system for intelligent robots Department of Electrical Engineering, Southern Taiwan University, Tainan County,
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Rank Minimization for Subspace Tracking from Incomplete Data
Robust Kernel Density Estimation by Scaling and Projection in Hilbert Space Presented by: Nacer Khalil.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Adaptive FIR Neural Model for Centroid Learning in Self-Organizing.
Scale Selection for Classification of Point-sampled 3-D Surfaces Jean-François Lalonde, Ranjith Unnikrishnan, Nicolas Vandapel and Martial Hebert Carnegie.
Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling The research leading to these results has received funding from the European.
Smooth ε -Insensitive Regression by Loss Symmetrization Ofer Dekel, Shai Shalev-Shwartz, Yoram Singer School of Computer Science and Engineering The Hebrew.
Non-negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs C.G. Puntonet and A. Prieto (Eds.): ICA 2004 Presenter.
NMF Demo: Lee, Seung Bryan Russell Computer Demonstration.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
CS 5751 Machine Learning Chapter 13 Reinforcement Learning1 Reinforcement Learning Control learning Control polices that choose optimal actions Q learning.

Distributed Computation Framework for Machine Learning
A Fast Trust Region Newton Method for Logistic Regression
Learning Mid-Level Features For Recognition
Recovery from Occlusion in Deep Feature Space for Face Recognition
Bryan Russell Computer Demonstration
Learning Incoherent Sparse and Low-Rank Patterns from Multiple Tasks
NON-NEGATIVE COMPONENT PARTS OF SOUND FOR CLASSIFICATION Yong-Choon Cho, Seungjin Choi, Sung-Yang Bang Wen-Yi Chu Department of Computer Science &
Learned Convolutional Sparse Coding
Presentation transcript:

Online Learning for Matrix Factorization and Sparse Coding Julien Mairal, Francis Bach, Jean Ponce and Guillermo Sapiro Journal of Machine Learning Research 2010

Introduction This paper focuses on the large scale matrix factorization problem, including Dictionary learning for sparse coding Non-negative matrix factorization (NMF) Sparse principal component analysis (SPCA) Contributions of this paper: An iterative online algorithm is proposed for large scale matrix factorization This algorithm is proved to converge almost surely to a stationary point of the objective function This algorithm is shown to be much faster than previous methods in the experiment.

Problem Statement Classical dictionary learning problem Given a finite training set , the objective is to optimize the following function where Online Learning This algorithm process one sample (or a mini-batch) at a time and sequentially minimize the following function:

Basic Algorithm

Dictionary Update

Optimizing the Algorithm Handling fixed-sized data sets Scaling the “past” data Mini-batch extension

Proof of Convergence Assumptions: Main results

Extensions to Matrix Factorization Non-negative matrix factorization (NMF) Non-negative sparse coding (NNSC) Sparse principal component analysis (SPCA)

Data for Experiment 1.25 million patches from Pascal VOC’06 image database

Online VS. Batch Training data size: 1 million OL1: OL2: OL3:

Comparison with NMF and NNSC

Face Results

Image Patches Results

Inpainting Results Image size: 12-Megapixel Dictionary with 256 elements Training data: 7 million 12 by 12 color patches

Conclusion A new online algorithm for learning dictionaries adapted to sparse coding tasks, and proven its convergence. Experiments demonstrate that this algorithm is significantly faster than existing batch methods. This algorithm can be extended to other matrix factorization problems such as non-negative matrix factorization and sparse principal component analysis.