Learning to Navigate for Fine-grained Classification

Slides:



Advertisements
Similar presentations
Adding Unlabeled Samples to Categories by Learned Attributes Jonghyun Choi Mohammad Rastegari Ali Farhadi Larry S. Davis PPT Modified By Elliot Crowley.
Advertisements

Foreground Focus: Finding Meaningful Features in Unlabeled Images Yong Jae Lee and Kristen Grauman University of Texas at Austin.
Parsing Clothing in Fashion Photographs
Learning on Probabilistic Labels Peng Peng, Raymond Chi-wing Wong, Philip S. Yu CSE, HKUST 1.
1 Building a Dictionary of Image Fragments Zicheng Liao Ali Farhadi Yang Wang Ian Endres David Forsyth Department of Computer Science, University of Illinois.
Robust Object Tracking via Sparsity-based Collaborative Model
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Unsupervised Transfer Classification Application to Text Categorization Tianbao Yang, Rong Jin, Anil Jain, Yang Zhou, Wei Tong Michigan State University.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Robust Higher Order Potentials For Enforcing Label Consistency
1 Integrating User Feedback Log into Relevance Feedback by Coupled SVM for Content-Based Image Retrieval 9-April, 2005 Steven C. H. Hoi *, Michael R. Lyu.
© 2013 IBM Corporation Efficient Multi-stage Image Classification for Mobile Sensing in Urban Environments Presented by Shashank Mujumdar IBM Research,
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
Nonparametric Part Transfer for Fine-grained Recognition Presenter Byungju Kim.
Learning Models for Object Recognition from Natural Language Descriptions Presenters: Sagardeep Mahapatra – Keerti Korrapati
Why Categorize in Computer Vision ?. Why Use Categories? People love categories!
Detection, Segmentation and Fine-grained Localization
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.
Reading Between The Lines: Object Localization Using Implicit Cues from Image Tags Sung Ju Hwang and Kristen Grauman University of Texas at Austin Jingnan.
1 Multiple Classifier Based on Fuzzy C-Means for a Flower Image Retrieval Keita Fukuda, Tetsuya Takiguchi, Yasuo Ariki Graduate School of Engineering,
BING: Binarized Normed Gradients for Objectness Estimation at 300fps
INTRODUCTION Heesoo Myeong and Kyoung Mu Lee Department of ECE, ASRI, Seoul National University, Seoul, Korea Tensor-based High-order.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Object detection, deep learning, and R-CNNs
Category Independent Region Proposals Ian Endres and Derek Hoiem University of Illinois at Urbana-Champaign.
Is Top-k Sufficient for Ranking? Yanyan Lan, Shuzi Niu, Jiafeng Guo, Xueqi Cheng Institute of Computing Technology, Chinese Academy of Sciences.
Vehicle Detection in Aerial Surveillance Using Dynamic Bayesian Networks Hsu-Yung Cheng, Member, IEEE, Chih-Chia Weng, and Yi-Ying Chen IEEE TRANSACTIONS.
Object Recognition as Ranking Holistic Figure-Ground Hypotheses Fuxin Li and Joao Carreira and Cristian Sminchisescu 1.
This tutorial will describe how to navigate the section of Gramene that allows you to view various types of maps (e.g., genetic, physical, or sequence-based)
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
When deep learning meets object detection: Introduction to two technologies: SSD and YOLO Wenchi Ma.
Big data classification using neural network
Object Detection based on Segment Masks
Compact Bilinear Pooling
Krishna Kumar Singh, Yong Jae Lee University of California, Davis
Saliency-guided Video Classification via Adaptively weighted learning
Jure Zbontar, Yann LeCun
Evaluating Techniques for Image Classification
A Pool of Deep Models for Event Recognition
Correlative Multi-Label Multi-Instance Image Annotation
Nonparametric Semantic Segmentation
ICCV Hierarchical Part Matching for Fine-Grained Image Classification
Structure-measure: A New Way to Evaluate Foreground Maps
Analysis of Sport Performance
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Location Recommendation — for Out-of-Town Users in Location-Based Social Network Yina Meng.
Bird-species Recognition Using Convolutional Neural Network
IEEE ICIP Feature Normalization for Part-Based Image Classification
Image Classification.
Multiple Instance Learning: applications to computer vision
Deep neural networks (DNNs) can successfully identify, count, and describe animals in camera-trap images. Deep neural networks (DNNs) can successfully.
“The Truth About Cats And Dogs”
Computational Imaging and Display Project Title
Open-Category Classification by Adversarial Sample Generation
Road Traffic Sign Recognition
On-going research on Object Detection *Some modification after seminar
CornerNet: Detecting Objects as Paired Keypoints
Object Detection Creation from Scratch Samsung R&D Institute Ukraine
Outline Background Motivation Proposed Model Experimental Results
Deep Cross-media Knowledge Transfer
A Unified View of Multi-Label Performance Measures
Learning Object Context for Dense Captioning
Zhedong Zheng, Liang Zheng and Yi Yang
Towards an Unequivocal Representation of Actions
Feature Selective Anchor-Free Module for Single-Shot Object Detection
Object Detection Implementations
Point Set Representation for Object Detection and Beyond
Introduction Face detection and alignment are essential to many applications such as face recognition, facial expression recognition, age identification,
Presentation transcript:

Learning to Navigate for Fine-grained Classification Ze Yang, Peking University

Problem Fine‐grained classification aims at differentiating categories that are very similar. For instance, the subordinate classes of a common superior class. The subordinate classes are similar in appearance.

Examples Determine plant species, breed of dogs, identification of dishes.

Examples Clothing recognition and retrieval

Examples Product recognition, smart retail

Key points to fine-grained classification Categories are different, but share a common part structure. The key point to fine‐grained classification lies in accurately identifying informative regions in the image.

Our works Learning to Navigate for fine-grained classification ECCV 2018

Motivations Intrinsic consistency between informativeness of the regions and their probability being ground‐truth class For informative regions, they will be assigned high probability being ground-truth class. But for uninformative regions that cannot help to differentiate classes, the classifier will not know their class and assigns them low probability being ground-truth class.

Overview Navigator: navigates the model to focus on informative regions. Teacher: evaluates the regions and provides feedback. Scrutinizer: scrutinizes those regions to make predictions.

Methodology Train the Navigator to propose informative regions. Navigator network is a RPN to compute the informativeness of all regions. We choose top‐M (here M=3) informative regions with informativeness {I1, I2, I3}. Then the Teacher network compute their confidences being GT class {C1, C2, C3}. We use ranking loss to optimize Navigator network to make {I1, I2, I3} and {C1, C2, C3} having the same order (function f is non‐decreasing). where the function f is a non-increasing function that encourages if Ranking loss:

Methodology The Scrutinizer makes predictions. Navigator network proposes the top‐K (here K=3) informative regions. Then the Scrutinizer network uses these regions and full image to make predictions. We use cross entropy loss to optimize the Teacher and the Scrutinizer.

Methodology Algorithm overview.

Experiments Quantitative results. Experimental results in CUB‐200‐2011. The table shows the comparison between our results and previous best results in CUB‐200‐2011. We use M=6 casually, which means top‐6 informative regions are used to train the Navigator. We also study the role of hyper‐parameter K, i.e. how many part regions have been used for fine‐grained classification.

Experiments Quantitative results. Experimental results in FGVA Aircraft and Stanford Cars datasets.

Experiments Qualitative results. The most informative regions proposed by Navigator network. We can see that the most informative regions are consistent with the human perception Birds: head, wings and main body Cars: headlamps and grilles Airplanes: wings and heads Especially in the blue box picture where the color of the bird and the background is quite similar.

Thank you