cs638/838 - Spring 2017 (Shavlik©), Week 10

Slides:



Advertisements
Similar presentations
Imbalanced data David Kauchak CS 451 – Fall 2013.
Advertisements

Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Ensemble Learning: An Introduction
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
How to bake a cake Or write a research essay. How many times have you felt like Sponge Bob?
Today’s Topics HW0 due 11:55pm tonight and no later than next Tuesday HW1 out on class home page; discussion page in MoodleHW1discussion page Please do.
CS Fall 2015 (© Jude Shavlik), Lecture 7, Week 3
Lecture 03 Area Based Image Processing Lecture 03 Area Based Image Processing Mata kuliah: T Computer Vision Tahun: 2010.
Today’s Topics Read –For exam: Chapter 13 of textbook –Not on exam: Sections & Genetic Algorithms (GAs) –Mutation –Crossover –Fitness-proportional.
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Today’s Topics HW1 Due 11:55pm Today (no later than next Tuesday) HW2 Out, Due in Two Weeks Next Week We’ll Discuss the Make-Up Midterm Be Sure to Check.
Today’s Topics Read: Chapters 7, 8, and 9 on Logical Representation and Reasoning HW3 due at 11:55pm THURS (ditto for your Nannon Tourney Entry) Recipe.
LECTURE 02: EVALUATING MODELS January 27, 2016 SDS 293 Machine Learning.
ImageNet Classification with Deep Convolutional Neural Networks Presenter: Weicong Chen.
Today’s Topics 11/10/15CS Fall 2015 (Shavlik©), Lecture 21, Week 101 More on DEEP ANNs –Convolution –Max Pooling –Drop Out Final ANN Wrapup FYI:
Artificial Neural Networks This is lecture 15 of the module `Biologically Inspired Computing’ An introduction to Artificial Neural Networks.
Analyzing the Behavior of Deep Models Dhruv Batra Devi Parikh Aishwarya Agrawal (EMNLP 2016)
Stock Market Application: Review
Welcome deep loria !.
Presented by Yuting Liu
cs638/838 - Spring 2017 (Shavlik©), Week 11
CS 636/838: BUILDING Deep Neural Networks
CS Fall 2016 (Shavlik©), Lecture 5
Deep learning David Kauchak CS158 – Fall 2016.
Artificial Intelligence
CS240A Final Project 2.
Ensembles (Bagging, Boosting, and all that)
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Why it is Called Tensor Flow Parallelism in ANNs Project Ideas and Discussion Glenn Fung Presents Batch Renormalizating Paper.
Computer Science and Engineering, Seoul National University
CS Fall 2016 (Shavlik©), Lecture 11, Week 6
cs540 - Fall 2015 (Shavlik©), Lecture 25, Week 14
Perceptrons Lirong Xia.
Neural Networks CS 446 Machine Learning.
Classification with Perceptrons Reading:
Supervised Learning Seminar Social Media Mining University UC3M
CS Fall 2016 (Shavlik©), Lecture 12, Week 6
Supervised Training of Deep Networks
CS 698 | Current Topics in Data Science
Google Analytics.
cs638/838 - Spring 2017 (Shavlik©), Week 7
CS Fall 2016 (© Jude Shavlik), Lecture 6, Week 4
Machine Learning Today: Reading: Maria Florina Balcan
Data Mining Practical Machine Learning Tools and Techniques
CSSE463: Image Recognition Day 11
Hyperparameters, bias-variance tradeoff, validation
CS 4501: Introduction to Computer Vision Training Neural Networks II
Logistic Regression & Parallel SGD
cs540 - Fall 2016 (Shavlik©), Lecture 20, Week 11
cs540 - Fall 2016 (Shavlik©), Lecture 18, Week 10
Lecture 18: Bagging and Boosting
Data Mining (and machine learning)
Neural Networks Geoff Hulten.
Lecture: Deep Convolutional Neural Networks
Ensemble learning.
CS Fall 2016 (Shavlik©), Lecture 12, Week 6
Tuning CNN: Tips & Tricks
Lecture 06: Bagging and Boosting
cs638/838 - Spring 2017 (Shavlik©), Week 7
Inception-v4, Inception-ResNet and the Impact of
Non-Symbolic AI lecture 9
Sketch Object Prediction
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Neural Machine Translation using CNN
Perceptrons Lirong Xia.
More on Maxent Env. Variable importance:
Sign Language Recognition With Unsupervised Feature Learning
Ensembles (Bagging, Boosting, and all that)
Presentation transcript:

cs638/838 - Spring 2017 (Shavlik©), Week 10 CS 540 Fall 2015 (Shavlik) 6/3/2018 Today’s Topics Some Lab 3 Comments Talk by Akshay Sood on recurrent ANNs, LSTM, etc Talk by Luisa Polania Cabrera of American Family Insurance on some of their Deep ML projects One-on-one Q&A on Lab3, Projects, etc I have run turned-in Lab 3 code Don’t redefine Vector! Project report comments emailed – lots on RL! April 18: Google-Madison talk on TPUs 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Sample, Latest CNN Results (32x32 images, batch size = 10) ADAM? drHU/drIn Train/TestErr@Epoch ExtraTrain #FlatHUs yes 0.00/0.00 0@20 18@20 18,636 256 yes 0.00/0.05 177@40 19@13 28,509 128 yes 0.50/0.05 113@80 20@74 18,568 256 no 0.00/0.00 410@30 22@25 28,460 256 no 0.00/0.00 0@160 25@159 0 64 ADAM worked well Extra examples helped a good deal Convolution kernels of 4x4 worked better than 5x5 (3x3 in between) ‘Zero padding’ worked about the same as w/o it Using 10 plates instead of 20 worked ok Better test-set accuracy than I expected, given small dataset! 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

cs638/838 - Spring 2017 (Shavlik©), Week 10 Ensembles Ensembles often greatly increase accuracy Combining all models with 25 or fewer testset errors, led to 12 errors! But this is a cheat! Why? Correcting the cheat, led to 19 errors Key question: how to pick N best models? Don’t forget about ensembles! 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Thought These are FLOWERS 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Thought These are AIRPLANES Some white borders! Not centered 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

cs638/838 - Spring 2017 (Shavlik©), Week 10 NOT butterflies More Ensemble Errors NOT pianos NOT starfish NOT a watch 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Tuneset vs. Testset Accuracies We’d like to threshold the Y axis, but we need to threshold the X ! Testeset Errors Tuneset Errors 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Some Lab 3 Report Comments Two senses of ‘learning curve’ (see original Lab 3 slides) CURVES better than TABLES! Some learning curves STEEP (next slide) suggests value in getting more original images Drop out worked for some, not for others Generating ‘perturbed’ examples greatly helps We really should replicate to get ‘error bars’ (ie, different random seeds) 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

An Encouraging Learning Curve! 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

My Learning Curve (used top-10 TUNEset models) Testeset Errors Number of (Original) Training Examples 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

cs638/838 - Spring 2017 (Shavlik©), Week 10 What Action is This? 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

cs638/838 - Spring 2017 (Shavlik©), Week 10 Impact of Random Seed Be careful to avoid ‘cherry picking’! Avoid ‘peeking at the test set’ while making decisions! 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

A Nice Overfitting Curve (from Lab 2) ERROR EPOCH 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10

Some Lab 3 Report Comments (2) Should start learning curves at 0 or ‘accuracy of always guessing most common output’ Some initial weights should be negative Too many plates might overfit? Third CONVOLUTION layer probably hurts WATCH predicted a lot because most common, probably not because it was LAST (reorder the enum!) ‘Rotated by 180’ – three meanings! 3/28/17 cs638/838 - Spring 2017 (Shavlik©), Week 10