Do Better ImageNet Models Transfer Better?

Slides:



Advertisements
Similar presentations
A brief review of non-neural-network approaches to deep learning
Advertisements

Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Limin Wang, Yu Qiao, and Xiaoou Tang
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Large-Scale Object Recognition with Weak Supervision
Visual Expertise Is a General Skill Maki Sugimoto University of California, San Diego November 20, 2000.
Spatial Pyramid Pooling in Deep Convolutional
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
A shallow introduction to Deep Learning
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Today Ensemble Methods. Recap of the course. Classifier Fusion
BEHAVIORAL TARGETING IN ON-LINE ADVERTISING: AN EMPIRICAL STUDY AUTHORS: JOANNA JAWORSKA MARCIN SYDOW IN DEFENSE: XILING SUN & ARINDAM PAUL.
Hierarchical Matching with Side Information for Image Classification
Locally Linear Support Vector Machines Ľubor Ladický Philip H.S. Torr.
1 Bernard Ng 1, Arash Vahdat 2, Ghassan Hamarneh 3, Rafeef Abugharbieh 1 Contact 1 Biomedical Signal and Image Computing Lab,
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Ke (Kevin) Wu1,2, Philip Watters1, Malik Magdon-Ismail1
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Wenchi MA CV Group EECS,KU 03/20/2017
Big data classification using neural network
Jonatas Wehrmann, Willian Becker, Henry E. L. Cagnini, and Rodrigo C
CS 4501: Introduction to Computer Vision Object Localization, Detection, Semantic Segmentation Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy.
Automatic Lung Cancer Diagnosis from CT Scans (Week 1)
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
The Relationship between Deep Learning and Brain Function
Compact Bilinear Pooling
Data Mining, Neural Network and Genetic Programming
Chilimbi, et al. (2014) Microsoft Research
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Understanding and Predicting Image Memorability at a Large Scale
Learning Mid-Level Features For Recognition
Article Review Todd Hricik.
Ajita Rattani and Reza Derakhshani,
Hierarchical Deep Convolutional Neural Network
Machine Learning Basics
Schizophrenia Classification Using
Deep learning and applications to Natural language processing
Training Techniques for Deep Neural Networks
Wide & deep networks.
Multiple Organ Detection in CT Volumes using CNN Week 4
Convolutional Neural Networks for sentence classification
Quanzeng You, Jiebo Luo, Hailin Jin and Jianchao Yang
Very Deep Convolutional Networks for Large-Scale Image Recognition
Pose Estimation for non-cooperative Spacecraft Rendevous using CNN
Lecture: Deep Convolutional Neural Networks
Outline Background Motivation Proposed Model Experimental Results
Deep Cross-media Knowledge Transfer
Biased Random Walk based Social Regularization for Word Embeddings
Tuning CNN: Tips & Tricks
Object Tracking: Comparison of
Designing Neural Network Architectures Using Reinforcement Learning
Logistic Regression & Transfer Learning
Ladislav Rampasek, Anna Goldenberg  Cell 
Heterogeneous convolutional neural networks for visual recognition
Abnormally Detection
Deep Interest Network for Click-Through Rate Prediction
Reuben Feinman Research advised by Brenden Lake
Car Damage Classification
Human-object interaction
Visual Question Answering
Presented by: Anurag Paul
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
An Evaluation Of Transfer Learning for Classifying Sales Engagement s at Large Scale Yong Liu, Pavel Dmitriev, Yifei Huang, Andrew Brooks, Li Dong.
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Learning and Memorization
Week 3 Volodymyr Bobyr.
Adrian E. Gonzalez , David Parra Department of Computer Science
Presentation transcript:

Do Better ImageNet Models Transfer Better? CVPR 2019 Oral Session 1-2C: Scenes & Representation Simon Kornblith, Jonathon Shlens, and Quoc V. Le Google Brain The last author is PhD Student of Andrew Ng.

Do better ImageNet models transfer better? Motivating Question Do better ImageNet models transfer better? How is the transferability of both ImageNet features and ImageNet classification architectures? A large-scale empirical study that systematically explores the problem.

Evaluation Datasets Networks Metrics Settings

12 Datasets

16 Models

Metrics accuracy: the meaning of a 1% additive increase in accuracy is different if it is relative to a base accuracy of 50% vs. 99%. logit-transformed accuracy: correlation: PLCC between the logit-transformed ImageNet accuracy and the logit-transformed transfer accuracy averaged across the 12 datasets

3 Settings training a logistic regression classifier on the fixed feature representation from the ImageNet-pretrained network fine-tuning the ImageNet-pretrained network training the same CNN architecture from scratch on the new image task.

The First Setting training a logistic regression classifier on the fixed feature representation from the ImageNet-pretrained network fine-tuning the ImageNet-pretrained network training the same CNN architecture from scratch on the new image task.

Results … when using an ImageNet-pretrained network as fixed feature extractor Green ones, not statistically different p-value 0.05 Permutation test for comparison on the same dataset t-test for comparison cross datasets better ImageNet architectures are capable of learning better, transferable representations.

ImageNet training settings affect transfer of fixed features

ImageNet training settings affect transfer of fixed features Some widely-used regularizers that improve ImageNet performance do not produce better representations. Low dimensional embeddings of Oxford 102 Flowers using t-SNE on features from Inception v4, for 10 classes from the test set.

The Second Setting training a logistic regression classifier on the fixed feature representation from the ImageNet-pretrained network fine-tuning the ImageNet-pretrained network training the same CNN architecture from scratch on the new image task.

Results … when fine-tuning an ImageNet-pretrained network

ImageNet training settings have only a minor impact on fine-tuning performance

The Third Setting training a logistic regression classifier on the fixed feature representation from the ImageNet-pretrained network fine-tuning the ImageNet-pretrained network training the same CNN architecture from scratch on the new image task.

Results … when training from scratch whether ImageNet accuracy for transfer learning is due to the weights derived from the ImageNet training or the architecture itself. 7 datasets (samples<10000) r=0.29 Other datasets r=0.86

Other Analysis Benefits of better models are comparable to specialized methods for transfer learning ImageNet pretraining does not necessarily improve accuracy on fine- grained tasks ImageNet pretraining accelerates convergence Accuracy benefits of ImageNet pretraining fade quickly with dataset size

Benefits of better models are comparable to specialized methods for transfer learning

ImageNet pretraining does not necessarily improve accuracy on fine-grained tasks

ImageNet pretraining accelerates convergence

Accuracy benefits of ImageNet pretraining fade quickly with dataset size

Conclusion and Comment A large-scale empirical study concludes that better ImageNet networks provide better features for transfer learning with linear classification, and better performance when the entire network is fine- tuned some regularizers that improve ImageNet performance are highly detrimental to the performance of transfer learning based on fixed features architectures transfer well across tasks even when weights do not This kind of research, i.e., systematic and deep analysis of the existing research, sometimes is even more beneficial to the research community than ``simply proposing a novel method``. Further readings Recht, Benjamin, et al. "Do ImageNet Classifiers Generalize to ImageNet?." International Conference on Machine Learning. 2019. On two small fine-grained classification datasets, fine-tuning does not provide a substantial benefit over training from random initialization, but better ImageNet architectures nonetheless obtain higher accuracy.