Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Using Large-Scale Web Data to Facilitate Textual Query Based Retrieval of Consumer Photos.
Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
Various Regularization Methods in Computer Vision Min-Gyu Park Computer Vision Lab. School of Information and Communications GIST.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
A CTION R ECOGNITION FROM V IDEO U SING F EATURE C OVARIANCE M ATRICES Kai Guo, Prakash Ishwar, Senior Member, IEEE, and Janusz Konrad, Fellow, IEEE.
SVM—Support Vector Machines
Support vector machine
Machine learning continued Image source:
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Large Scale Manifold Transduction Michael Karlen Jason Weston Ayse Erkan Ronan Collobert ICML 2008.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
More MR Fingerprinting
Ilias Theodorakopoulos PhD Candidate
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Large-Scale, Real-World Face Recognition in Movie Trailers Week 2-3 Alan Wright (Facial Recog. pictures taken from Enrique Gortez)
Classification and Decision Boundaries
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
Principal Component Analysis
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
ACM Multimedia th Annual Conference, October , 2004
An Introduction to Kernel-Based Learning Algorithms K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda and B. Scholkopf Presented by: Joanna Giforos CS8980: Topics.
Reduced Support Vector Machine
Support Vector Machines Formulation  Solve the quadratic program for some : min s. t.,, denotes where or membership.  Different error functions and measures.
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Linear Discriminant Functions Chapter 5 (Duda et al.)
Optimizing Learning with SVM Constraint for Content-based Image Retrieval* Steven C.H. Hoi 1th March, 2004 *Note: The copyright of the presentation material.
Ch. Eick: Support Vector Machines: The Main Ideas Reading Material Support Vector Machines: 1.Textbook 2. First 3 columns of Smola/Schönkopf article on.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Machine Learning Week 4 Lecture 1. Hand In Data Is coming online later today. I keep test set with approx test images That will be your real test.
This week: overview on pattern recognition (related to machine learning)
Active Learning for Class Imbalance Problem
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Machine Learning Seminar: Support Vector Regression Presented by: Heng Ji 10/08/03.
计算机学院 计算感知 Support Vector Machines. 2 University of Texas at Austin Machine Learning Group 计算感知 计算机学院 Perceptron Revisited: Linear Separators Binary classification.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Classifiers Given a feature representation for images, how do we learn a model for distinguishing features from different classes? Zebra Non-zebra Decision.
Applications of Neural Networks in Time-Series Analysis Adam Maus Computer Science Department Mentor: Doctor Sprott Physics Department.
Transductive Regression Piloted by Inter-Manifold Relations.
D. M. J. Tax and R. P. W. Duin. Presented by Mihajlo Grbovic Support Vector Data Description.
Face Detection Using Large Margin Classifiers Ming-Hsuan Yang Dan Roth Narendra Ahuja Presented by Kiang “Sean” Zhou Beckman Institute University of Illinois.
© 2009 IBM Corporation IBM Research Xianglong Liu 1, Yadong Mu 2, Bo Lang 1 and Shih-Fu Chang 2 1 Beihang University, Beijing, China 2 Columbia University,
CS 1699: Intro to Computer Vision Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh October 29, 2015.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Unsupervised Auxiliary Visual Words Discovery for Large-Scale Image Object Retrieval Yin-Hsi Kuo1,2, Hsuan-Tien Lin 1, Wen-Huang Cheng 2, Yi-Hsuan Yang.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Classification Course web page: vision.cis.udel.edu/~cv May 14, 2003  Lecture 34.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Motion Segmentation CAGD&CG Seminar Wanqiang Shen
Cross-modal Hashing Through Ranking Subspace Learning
1 Bilinear Classifiers for Visual Recognition Computational Vision Lab. University of California Irvine To be presented in NIPS 2009 Hamed Pirsiavash Deva.
Day 17: Duality and Nonlinear SVM Kristin P. Bennett Mathematical Sciences Department Rensselaer Polytechnic Institute.
Linear Discriminant Functions Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Deeply learned face representations are sparse, selective, and robust
Intelligent Information System Lab
Pawan Lingras and Cory Butz
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
CS 2750: Machine Learning Support Vector Machines
COSC 4335: Other Classification Techniques
Optimal sparse representations in general overcomplete bases
Connecting Data with Domain Knowledge in Neural Networks -- Use Deep learning in Conventional problems Lizhong Zheng.
Support Vector Machines
Outline Sparse Reconstruction RIP Condition
Presentation transcript:

Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign

 Motivation  Sparse-Graph based Semi-supervised Learning  Handling of Noisy Tags  Inferring Concepts in Semantic Concept Space  Experiments  Summarization and Future Work

No manual annotation are required.

 With models:  SVM  GMM  …  Infer labels directly:  k-NN  Graph-based semi-supervised methods

 A common disadvantage:  Have certain parameters that require manual tuning  Performance is sensitive to parameter tuning  The graphs are constructed based on visual distance  Many links between samples with unrelated-concepts  The label information will be propagated incorrectly.  Locally linear reconstruction:  Still needs to select neighbors based on visual distance

 Sparse Graph based Learning  Noisy Tag Handling  Inferring Concepts in the Concept Space

 Human vision system seeks a sparse representation for the incoming image using a few visual words in a feature vocabulary. ( Neural Science )  Advantages:  Reduce the concept-unrelated links to avoid the propagation of incorrect information;  Practical for large-scale applications, since the sparse representation can reduce the storage requirement and is feasible for large-scale numerical computation.

Normal Graph Construction. Sparse Graph Construction.

 The ℓ 1 -norm based linear reconstruction error minimization can naturally lead to a sparse representation for the images *. * J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma. Robust face recognition via sparse representation. IEEE Transaction on Pattern Analysis and Machine Intelligence, 31(2):210– 227, Feb  The sparse reconstruction can be obtained by solving the following convex optimization problem: min w ||w|| 1, s.t. x=Dw w ∈ R n : the vector of the reconstruction coefficients; x ∈ R d : feature vector of the image to be reconstructed; D ∈ R d*n (d < n) : a matrix formed by the feature vectors of the other images in the dataset.

 Handle the noise on certain elements of x:  Reformulate x = Dw+ξ, where ξ ∈ R d is the noise term.  Then :  Set the edge weight of the sparse graph:

 Result:

 The problem with :  M uu is typically very large for image annotation  It is often computationally prohibitive to calculate its inverse directly  Iterative solution with non-negative constraints: may not be reasonable since some samples may have negative contributions to the other samples  Solution:  Reformulate: The generalized minimum residual method (usually abbreviated as GMRES) can be used to iteratively solve this large-scale sparse system of linear equations effectively and efficiently.

√: correct; ?: ambiguous; m: missing

 We cannot assume that the training tags are fixed during the inference process.  The noisy training tags should be refined during the label inference.  Solution: adding two regularization terms into the inferring framework to handle the noise:

 Solution:  Set the original label vector as the initial estimation of ideal label vector, that is, set, and then solve and we can obtain a refined f l.  Fix f l and solve  Use the obtained to replace the y in the previous graph-based method, and we can solve the sparse system of linear equations to infer the labels of the unlabeled samples.

 It is well-known that inferring concepts based on low-level visual features cannot work very well due to the semantic gap.  To bridge this semantic gap  Construct a concept space and then infer the semantic concepts in this space.  The semantic relations among different concepts are inherently embedded in this space to help the concept inference.

 Low-semantic-gap: Concepts in the constructed space should have small semantic gaps;  Informative: These concepts can cover the semantic space spanned by all useful concepts (tags), that is, the concept space should be informative;  Compact: The set including all the concepts forming the space should be compact (i.e., the dimension of the concept space is small).

 Basic terms:  Ω : the set of all concepts;  Θ : the constructed concept set.  Three measures:  Semantic Modelability: SM(Θ)  Coverage of Semantic Concept Space: CE(Θ, Ω)  Compactness: CP(Θ)=1/#(Θ)  Objective:

 Simplification: fix the size of the concept space.  Then we can transform this maximization to a standard quadratic programming problem.  See the paper for more details.

 Image mapping: x i D(i)  Query concept mapping: c x Q(c x )  Ranking the given images:

 Dataset  NUS-WIDE Lite Version (55,615 images)  Low-level Features  Color Histogram (CH) and Edge Direction Histogram (EDH), combine directly.  Evaluation  81 concepts  AP and MAP

Ex1: Comparisons among Different Learning Methods

 Ex2: Concept Inference with and without Concept Space

Ex3: Inference with Tags vs. Inference with Ground-truth We can achieve an MAP of by inference from tags in the concept space, which is comparable to the MAP obtained by inference from ground-truth of training labels.

 Exploited the problem of inferring semantic concepts from community-contributed images and their associated noisy tags.  Three points:  Sparse graph based label propagation  Noisy tag handling  Inference in a low-semantic-gap concept space

 Training set construction from the web resource

Thanks! Questions?