Chapter 11 Practical Methodology

Slides:



Advertisements
Similar presentations
Classification Classification Examples
Advertisements

Imbalanced data David Kauchak CS 451 – Fall 2013.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Deep Learning and Neural Nets Spring 2015
Introduction to Machine Learning Approach Lecture 5.
Radial Basis Function Networks
November 25, 2014Computer Vision Lecture 20: Object Recognition IV 1 Creating Data Representations The problem with some data representations is that the.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Deep Learning and Neural Nets Spring 2015
Using Neural Networks in Database Mining Tino Jimenez CS157B MW 9-10:15 February 19, 2009.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
NEURAL NETWORKS FOR DATA MINING
Machine learning system design Prioritizing what to work on
CSSE463: Image Recognition Day 11 Lab 4 (shape) tomorrow: feel free to start in advance Lab 4 (shape) tomorrow: feel free to start in advance Test Monday.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
CSSE463: Image Recognition Day 11 Due: Due: Written assignment 1 tomorrow, 4:00 pm Written assignment 1 tomorrow, 4:00 pm Start thinking about term project.
1 CSI5388 Practical Recommendations. 2 Context for our Recommendations I This discussion will take place in the context of the following three questions:
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Naïve Bayes Classification Recitation, 1/25/07 Jonathan Huang.
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Chapter 12: Simulation and Modeling
OPERATING SYSTEMS CS 3502 Fall 2017
Machine Learning for Computer Security
Evaluating Classifiers
Deep Feedforward Networks
Deep Learning Amin Sobhani.
Signal processing and Networking for Big Data Applications Lecture 17: Deep Learning Methodology and Applications Zhu Han University of Houston Thanks.
PRESENTATION FOR 3rd DEGREE COMPULSORY SECONDARY EDUCATION STUDENTS
CSSE463: Image Recognition Day 11
Generative Adversarial Networks
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Perceptrons Lirong Xia.
Pattern Recognition Sergios Theodoridis Konstantinos Koutroumbas
Chapter 1 - An Introduction to Computers and Problem Solving
Data Mining Lecture 11.
CSC 578 Neural Networks and Deep Learning
CSSE463: Image Recognition Day 11
Introduction to Systems Analysis and Design
CS 4501: Introduction to Computer Vision Training Neural Networks II
Deep Learning Hierarchical Representations for Image Steganalysis
cs540 - Fall 2016 (Shavlik©), Lecture 20, Week 11
Discrete Event Simulation - 4
Chap. 7 Regularization for Deep Learning (7.8~7.12 )
Overview of Machine Learning
Practical Considerations
Computer Vision Chapter 4
Creating Data Representations
Getting Started BPS 7e Chapter 0 © 2014 W. H. Freeman and Company.
Presentation and project
Autoencoders hi shea autoencoders Sys-AI.
Neural networks (3) Regularization Autoencoder
Evaluating Classifiers
Computer Vision Lecture 19: Object Recognition III
CSSE463: Image Recognition Day 11
Roc curves By Vittoria Cozza, matr
CSSE463: Image Recognition Day 11
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Introduction to Neural Networks
NAÏVE BAYES CLASSIFICATION
Perceptrons Lirong Xia.
CSC 578 Neural Networks and Deep Learning
Patterson: Chap 1 A Review of Machine Learning
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Chapter 11 Practical Methodology 11.1 Performance Metrics 11.2 Default Baseline Models 11.3 Determining Whether to Gather More Data 11.4. Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning 전자전기컴퓨터공학부 공민영

11.1 Performance Metrics Determine your goals You will also be limited by having a finite amount of training data. Data collection can require time, money, or human suffering. How can one determine a reasonable level of performance to expect? determined desired error rate your design decisions will be guided by reaching this error rate.

11.1 Performance Metrics The accuracy, or equivalently, the error rate, of a system. However, many applications require more advanced metrics. Ex) An e-mail spam detection system can make two kinds of mistakes: incorrectly classifying a legitimate message as spam, > incorrectly allowing a spam message to appear in the in box.

11.1 Performance Metrics Sometimes we wish to train a binary classifier that is intended to detect some rare event. We can easily achieve 99.9999% accuracy on the detection task, by simply hard-coding the classifier to always report that the disease is absent. Clearly, accuracy is a poor way to characterize the performance of such a system.

11.1 Performance Metrics Precision and recall. Precision the fraction of detections reported by the model that were correct, recall the fraction of true events that were detected.

11.1 Performance Metrics PR curve(Precision-Recall curve) precision on they-axis and recall on the x-axis Ex) to detect a disease outputs Medical result The probability that a person has the disease. We choose to report a detection whenever this score exceeds some threshold. By varying the threshold, we can trade precision for recall. F-score (precision p and recall r)

11.2 Default Baseline Models establish a reasonable end-to-end system. supervised learning with fixed-size vectors ->feedforward network input has Known topological structure(eg image) -> convolutional network Batch normalization can have a dramatic effect on optimization performance. Early stopping should be used almost universally. Dropout is an excellent regularizer that is easy to implement and compatible with many models and training algorithms. Batch normalization also sometimes reduces generalization error and allows dropout to be omitted, due to the noise in the estimate of the statistics used to normalize each variable.

11.2 Default Baseline Models it is common to use the features from a convolutional network trained on ImageNet to solve other computer vision tasks. unsupervised learning is somewhat domain specific. natural language processing are known to benefit tremendously from unsupervised learning. such as computer vision, unsupervised learning techniques do not bring a benefit. If your application is in a context where unsupervised learning is known to be important, then include it in your first end-to-end baseline. You can always try adding unsupervised learning later if you observe that your initial baseline overfits.

11.3 Determining Whether to Gather More Data YES Measure the performance on training set Acceptable? NO No more data Increase the model YES Acceptable? NO More training data YES Measure the performance on test set

Measure the performance on test set 11.3 Determining Whether to Gather More Data Measure the performance on test set Acceptable? NO Gathering more data YES NO Acceptable? YES Done

11.3 Determining Whether to Gather More Data Gathering more data the cost and feasibility of gathering more data. the cost and feasibility of reducing the test error by other means amount of data that is expected to be necessary to improve test set the answer is almost always to gather more training data.

11.3 Determining Whether to Gather More Data To decide how much to gather It is recommended to experiment with training set sizes on a logarithmic scale, for example doubling the number of examples between consecutive experiments

11.4 Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning Hyperparameters the representational capacity effective capacity The ability of the learning algorithm the degree to which the cost function and training procedure regularize the model.

11.4 Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning

11.4 Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning learning rate is important. If you have time to tune only one hyperparameter, tune the learning rate. It controls the effective capacity of the model in a more complicated way than other hyperparameters.

11.4 Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning

11.4 Selecting Hyperparameters 11.4.1 Manual Hyperparameter Tuning Training error > target error -> increase capacity(more layer or more hidden units.. But cost↑) Test error > target error Test error = training error + the gap between train and test error training error is very low(when best performance in Neural network) Test error ≒ the gap between train and test error Usually the best performance comes from a large model that is regularized well, for example by using dropout.

11.4 Selecting Hyperparameters