Thesis Advisor : Prof C.V. Jawahar

Slides:



Advertisements
Similar presentations
Attributes for Classifier Feedback Amar Parkash and Devi Parikh.
Advertisements

Active Shape Models Suppose we have a statistical shape model –Trained from sets of examples How do we use it to interpret new images? Use an “Active Shape.
Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
Recognizing Human Actions by Attributes CVPR2011 Jingen Liu, Benjamin Kuipers, Silvio Savarese Dept. of Electrical Engineering and Computer Science University.
Random Forest Predrag Radenković 3237/10
Presented by Xinyu Chang
Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Pattern Recognition and Machine Learning
Face Alignment with Part-Based Modeling
« هو اللطیف » By : Atefe Malek. khatabi Spring 90.
Learning Visual Similarity Measures for Comparing Never Seen Objects Eric Nowak, Frédéric Jurie CVPR 2007.
Mixture of trees model: Face Detection, Pose Estimation and Landmark Localization Presenter: Zhang Li.
Face Detection, Pose Estimation, and Landmark Localization in the Wild
Intelligent Systems Lab. Recognizing Human actions from Still Images with Latent Poses Authors: Weilong Yang, Yang Wang, and Greg Mori Simon Fraser University,
Human Action Recognition by Learning Bases of Action Attributes and Parts.
Discriminative Segment Annotation in Weakly Labeled Video Kevin Tang, Rahul Sukthankar Appeared in CVPR 2013 (Oral)
Content Based Image Clustering and Image Retrieval Using Multiple Instance Learning Using Multiple Instance Learning Xin Chen Advisor: Chengcui Zhang Department.
Relative Attributes Presenter: Shuai Zheng (Kyle) Supervised by Philip H.S. Torr Author: Devi Parikh (TTI-Chicago) and Kristen Grauman (UT-Austin)
Image Search Presented by: Samantha Mahindrakar Diti Gandhi.
Decision Tree Algorithm
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
Presented by Zeehasham Rasheed
5/30/2006EE 148, Spring Visual Categorization with Bags of Keypoints Gabriella Csurka Christopher R. Dance Lixin Fan Jutta Willamowski Cedric Bray.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Spatial Pyramid Pooling in Deep Convolutional
© 2013 IBM Corporation Efficient Multi-stage Image Classification for Mobile Sensing in Urban Environments Presented by Shashank Mujumdar IBM Research,
Ensemble Learning (2), Tree and Forest
Wang, Z., et al. Presented by: Kayla Henneman October 27, 2014 WHO IS HERE: LOCATION AWARE FACE RECOGNITION.
Exercise Session 10 – Image Categorization
Bag-of-Words based Image Classification Joost van de Weijer.
Identifying Computer Graphics Using HSV Model And Statistical Moments Of Characteristic Functions Xiao Cai, Yuewen Wang.
Computer vision.
A Statistical Approach to Speed Up Ranking/Re-Ranking Hong-Ming Chen Advisor: Professor Shih-Fu Chang.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Reading Between The Lines: Object Localization Using Implicit Cues from Image Tags Sung Ju Hwang and Kristen Grauman University of Texas at Austin Jingnan.
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Sharing Features Between Objects and Their Attributes Sung Ju Hwang 1, Fei Sha 2 and Kristen Grauman 1 1 University of Texas at Austin, 2 University of.
C. Lawrence Zitnick Microsoft Research, Redmond Devi Parikh Virginia Tech Bringing Semantics Into Focus Using Visual.
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Some Aspects of Bayesian Approach to Model Selection Vetrov Dmitry Dorodnicyn Computing Centre of RAS, Moscow.
First-Person Activity Recognition: What Are They Doing to Me? M. S. Ryoo and Larry Matthies Jet Propulsion Laboratory, California Institute of Technology,
ACADS-SVMConclusions Introduction CMU-MMAC Unsupervised and weakly-supervised discovery of events in video (and audio) Fernando De la Torre.
Dense Color Moment: A New Discriminative Color Descriptor Kylie Gorman, Mentor: Yang Zhang University of Central Florida I.Problem:  Create Robust Discriminative.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Hybrid Classifiers for Object Classification with a Rich Background M. Osadchy, D. Keren, and B. Fadida-Specktor, ECCV 2012 Computer Vision and Video Analysis.
Advanced Gene Selection Algorithms Designed for Microarray Datasets Limitation of current feature selection methods: –Ignores gene/gene interaction: single.
Detecting Remote Evolutionary Relationships among Proteins by Large-Scale Semantic Embedding Xu Linhe 14S
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Evaluation of Gender Classification Methods with Automatically Detected and Aligned Faces Speaker: Po-Kai Shen Advisor: Tsai-Rong Chang Date: 2010/6/14.
Zhuode Liu 2016/2/13 University of Texas at Austin CS 381V: Visual Recognition Discovering the Spatial Extent of Relative Attributes Xiao and Lee, ICCV.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Chapter 7. Classification and Prediction
Multimedia Content-Based Retrieval
Can Computer Algorithms Guess Your Age and Gender?
Hybrid Features based Gender Classification
Histogram—Representation of Color Feature in Image Processing Yang, Li
Accounting for the relative importance of objects in image retrieval
Face detection using Random projections
Attributes and Simile Classifiers for Face Verification
Learning with information of features
Ying shen Sse, tongji university Sep. 2016
Outline Background Motivation Proposed Model Experimental Results
Machine Learning: UNIT-4 CHAPTER-1
Zeroshot Learning Mun Jonghwan.
Presentation transcript:

Thesis Advisor : Prof C.V. Jawahar

What are Visual Attributes? Mid level concepts that bridge the gap between low level image features ( texture) and high level concepts (e.g: beach, table, Tom cruise). Examples are properties of materials (furry, metallic), faces ( young, female) etc. RED STRIPED

Furry dog Young Female Curly hair

Attribute properties Shareable across different but related concepts. e.g: (Red shirt, Red car), ( open street, open corridor) Both visual and semantic, ideal for communication. e.g: “ these red flowers are beautiful” Used in visual search.

What are relative attributes ? Binary attributes indicate only the presence and absence of attributes. Relative attributes introduced by Parikh and Grauman provides more appealing way of comparing two images

Relative Attributes Provides a semantic richer way to describe and compare objects in the world. Help in refining an identifying description or to situate with respect to reference objects ( brighter than a candle and dimmer than a light). D. Parikh and K. Grauman Relative Attributes, ICCV 2011

They help in visual search E.g: Find me black shoes which are more shinier than these and more formal than these.

Problem Statement In this Thesis, we concentrated on the problem of predicting relative attributes for facial domain. Given two images which contain faces, we aim to predict the relative strength of given attribute.

Pairs of images are used for training the model with their relative strengths Smiling attribute

During Testing model predicts the score for each image. Smiling attribute

Learning of relative attributes classifier For training of relative attribute classifiers, pair of images with their relative attribute strengths are required. For testing, given two images, classifier has to give high score for the image which has more strength of attribute than the other. We have used Ranking SVM classifier to learn the model for ranking images.

Distinction between Classification SVM and Ranking SVM Binary classifier tries to separate the positive and negative samples whereas Ranking SVM tries to order the points based on their strength.

Using parts for learning Relative attributes The visual attributes we have considered are more local in nature This intuition makes us to use a part based local representation rather than using a global feature representation. Part-to-Part comparison is effective and accurate. This helps to distinguish the fine changes in the local regions.

Detecting parts on face Parts on face need to be detected to compute part specific features. In our work, we have used the method of “Face detection, pose estimation and landmark localization in the wild CVPR 2012" which uses tree structure models. Example image where parts are detected using the above method. The upper parts of face are detected by extending the parts which are already detected.

Feature Representation Bag of words feature vector is computed for each part. Dense sift features are used for feature representation. Feature vector from each part are concatenated to form feature representation for entire image.

Overview of Our Method

Data Sets and Data Collection “Pubfig-29” dataset which contains attribute labels for 29 attributes. “Pubfig-29” contains class specific annotations rather than image level annotations. We found that image level annotations provide more detailed and accurate information than the class level annotations. We have collected a new dataset called “LFW-10” which contains attribute labels for 10 different attributes. Training set contains 500 pairs for each attribute. Testing set contains 500 pairs for each attribute.

LFW -10 dataset

Learning weights for parts Learn which parts are more important than the other. We assign significance coefficients for each part, which tells how significant is that part with respect to other parts in predicting the attribute.

Problem formulation

Solving Optimization Problem Solved through block co-ordinate descent algorithm. Consider each set of parameters Wm and Sm as two blocks, and optimize them in an alternate manner. All entries of Wm are initialized to be zero, and all entries of Sm are initialized to be equal to 1/K. ( K = no of parts) Alternatively learn the parameters Sm and Wm until accuracy on the validation set stops increasing.

Experiments We compare our proposed method to the method of Relative Attributes[ ICCV 2011] on LFW dataset under different settings. Features for parts: we represent a part as BOW histogram over dense SIFT features. We used two different settings for this 1) part specific vocabulary 2) Single vocabulary for all the parts. Baselines: we compare Ranking SVM method with different features. 1) BOW histogram over 1000 visual words. 2) Global 512 dimensional GIST feature vector over the entire image. 3) Global 512 dimensional GIST and 30 dimensional RGB feature vector over the entire image. 4) Spatial pyramid up to 2 or 3 levels using dense sift + BOW.

Results

Top 10 parts learned by our method for each attribute

Application to interactive image search Used to search an image by comparing with other images using relative attribute feedback. The user can search for a particular image by selecting images from the database and giving feedback

Results for Image search application Performance variation of different methods on interactive image search with number of reference images and number of feedbacks. Each plot shows the number of searches in which the target image is ranked below a particular rank. Larger is the number of searches falling below a specified rank, better is the accuracy

Relative attributes to relational attributes Relational Attributes provide a more natural way of comparing two images based on some given attribute than the Relative attribute. Relational attributes take into account not only the content of the two images but also its relation to the other image pairs. This makes the comparison more robust. Intuition behind this is when we use difference of features between images the contribution coming from the image pair will not be very significant when compared to contribution coming from all other image pairs.

Gaussian process regression In the method of relative attributes, ranking is based on the image pair. Only the content of two images is taken into consideration. In our method given two images to rank, we also consider the relationship of this pair with all other pairs using Gaussian process regression. Gaussian process is a collection of random variables such that each random variable corresponds to a Gaussian distribution. In our case each random variable is an image pair.

We define a pairwise ranking function as Where K is predefined kernel that gives a gram matrix. Xpq ( Xp-Xq) is the difference of features of two images. The above equation can be written as Gaussian kernel is used to compute similarity between two images. Optimized the problem using Newtons method. Fixed parameter σ = 0.3 and optimize γ using cross validation.

Experiments and results Dataset : Pub-fig Training : 3300 pairs, Testing : 1100 pairs. Five fold cross validation. Effect of training size on accuracy. We compare our method with Ranking SVM [Relative Attributes ICCV 2011]

Conclusion In this thesis, we have explored the problem of Relative attribute Prediction and presented part based solution to solve the problem. The Part based representation that we have proposed for learning relative attributes gives significant improvement in accuracy over the previous methods. Apart from the relative attribute prediction, we have shown advantage of our method on Interactive Image search. Inspired from the success of attributes and relative attributes, we present relational attributes. Despite simplicity of the approach it shows superior results than relative attributes.

References [1] Relative attributes D.Parikh, K.Grauman ICCV 2011 [2] Relative Parts: Distinctive Parts for Learning Relative Attributes. Ramachandruni N. Sandeep. Yashaswi Verma. C. V. Jawahar. CVPR 2014. [3] Our Project Page: http://cvit.iiit.ac.in/projects/relativeParts/

Thank you