Lecture 3: آشنایی با کتابخانه FastAi پیدا کردن Learning Rate مناسب

Slides:



Advertisements
Similar presentations
Optimization Tutorial
Advertisements

Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Lecture 29: Optimization and Neural Nets CS4670/5670: Computer Vision Kavita Bala Slides from Andrej Karpathy and Fei-Fei Li
Linear Regression  Using a linear function to interpolate the training set  The most popular criterion: Least squares approach  Given the training set:
The Widrow-Hoff Algorithm (Primal Form) Repeat: Until convergence criterion satisfied return: Given a training set and learning rate Initial:  Minimize.
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Lecture 4: CNN: Optimization Algorithms
Collaborative Filtering Matrix Factorization Approach
Mathematical formulation XIAO LIYING. Mathematical formulation.
Online Learning for Collaborative Filtering
Training begins in… 15:00 minutes Training begins in… 14:00 minutes.
Insight: Steal from Existing Supervised Learning Methods! Training = {X,Y} Error = target output – actual output.
Tips for Training Neural Network
Logistic Regression William Cohen.
Steepest Descent Method Contours are shown below.
Lecture 9 State Space Gradient Descent Gibbs Sampler with Simulated Annealing.
WEEK 2 SOFT COMPUTING & MACHINE LEARNING YOSI KRISTIAN Gradient Descent for Linear Regression.
From TOPIC to MAIN IDEA A process exercise. Go to the number that matches your number! 1. Read the article. Write the TOPIC! 2. Read the article. Fix.
Neural Networks - Berrin Yanıkoğlu1 MLP & Backpropagation Issues.
CSC321: Neural Networks Lecture 9: Speeding up the Learning
Matt Gormley Lecture 4 September 12, 2016
Multiplying Decimals.
Starter – Try this exam question! C2 June 2012
Supervised Learning in ANNs
Recurrent Neural Networks for Natural Language Processing
A Fast Trust Region Newton Method for Logistic Regression
Flexibility and Adaptability
CS623: Introduction to Computing with Neural Nets (lecture-5)
Machine Learning – Regression David Fenyő
Methodologies in Computing
Machine Learning I & II.
Neural Networks CS 446 Machine Learning.
Softmax Classifier + Generalization
Machine Learning – Regression David Fenyő
Neural Networks and Backpropagation
CSC 578 Neural Networks and Deep Learning
Face Recognition with Deep Learning Method
Disadvantages of Discrete Neurons
Differentiation.
Multiplying Decimals.
Collaborative Filtering Matrix Factorization Approach
Logistic Regression & Parallel SGD
Artificial Intelligence 13. Multi-Layer ANNs
cs540 - Fall 2016 (Shavlik©), Lecture 18, Week 10
Large Scale Support Vector Machines
Perceptron as one Type of Linear Discriminants
EET 2259 Unit 6 Shift Registers
Multilayer Perceptron & Backpropagation
Lecture: Image manipulations (2) Filters and kernels
Lecture 7: Simple Classifier (KNN)
Lecture 6: Introduction to Machine Learning
Convolutional networks
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
Softmax Classifier.
Lecture 2 CMS 165 Optimization.
Image Classification & Training of Neural Networks
CS623: Introduction to Computing with Neural Nets (lecture-5)
Crypto Encryption Intro to public key.
Which equation does the function {image} satisfy ?
Multiple features Linear Regression with multiple variables
Multiple features Linear Regression with multiple variables
Introduction to Neural Networks
SOL 8.8 Students will be using the 8.8 Transformation Chart for Notes
EET 2259 Unit 6 Shift Registers
CSC 578 Neural Networks and Deep Learning
Overall Introduction for the Lecture
REU Week 7: Real-Time Video Anomaly Detection
Presentation transcript:

Lecture 3: آشنایی با کتابخانه FastAi پیدا کردن Learning Rate مناسب Alireza Akhavan Pour CLASS.VISION شنبه، ۲۱ مهر ۱۳۹۷

How does learning rate impact training? شنبه، ۲۱ مهر ۱۳۹۷

یافتن LR مناسب https://arxiv.org/pdf/1506.01186.pdf شنبه، ۲۱ مهر ۱۳۹۷

یافتن LR مناسب fastest decrease in the loss شنبه، ۲۱ مهر ۱۳۹۷

https://arxiv.org/pdf/1608.03983.pdf شنبه، ۲۱ مهر ۱۳۹۷

Stochastic Gradient Descent with Restarts (SGDR) شنبه، ۲۱ مهر ۱۳۹۷

Stochastic Gradient Descent with Restarts (SGDR) The idea behind SGDR as shown in this image is, instead of trying to add various forms of learning rate decay, let’s reset our learning rate every so many iterations so that we may be able to more easily pop out of a local minimum if we appear stuck. This has seemed to be quite an improvement in various situations as compared to the normal SGD using mini batches. شنبه، ۲۱ مهر ۱۳۹۷

Stochastic Gradient Descent with Restarts (SGDR) شنبه، ۲۱ مهر ۱۳۹۷

Stochastic Gradient Descent with Restarts (SGDR) شنبه، ۲۱ مهر ۱۳۹۷

Transfer Learning using differential learning rates شنبه، ۲۱ مهر ۱۳۹۷

Transfer Learning using differential learning rates شنبه، ۲۱ مهر ۱۳۹۷

منابع https://www.coursera.org/specializations/deep-learning http://course.fast.ai/ https://medium.com/38th-street-studios/exploring-stochastic-gradient-descent-with-restarts-sgdr-fa206c38a74e https://towardsdatascience.com/transfer-learning-using-differential-learning-rates-638455797f00 شنبه، ۲۱ مهر ۱۳۹۷