Roberto Battiti, Mauro Brunato

Slides:



Advertisements
Similar presentations
Learning Algorithm Evaluation
Advertisements

R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Indian Statistical Institute Kolkata
Evaluation.
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
Evaluation.
Experimental Evaluation
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
Sampling : Error and bias. Sampling definitions  Sampling universe  Sampling frame  Sampling unit  Basic sampling unit or elementary unit  Sampling.
CLassification TESTING Testing classifier accuracy
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 5 of Data Mining by I. H. Witten, E. Frank and M. A. Hall 報告人:黃子齊
Outline 1-D regression Least-squares Regression Non-iterative Least-squares Regression Basis Functions Overfitting Validation 2.
Last lecture summary. Basic terminology tasks – classification – regression learner, algorithm – each has one or several parameters influencing its behavior.
CpSc 810: Machine Learning Evaluation of Classifier.
Manu Chandran. Outline Background and motivation Over view of techniques Cross validation Bootstrap method Setting up the problem Comparing AIC,BIC,Crossvalidation,Bootstrap.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
Assignments CS fall Assignment 1 due Generate the in silico data set of 2sin(1.5x)+ N (0,1) with 100 random values of x between.
Validation methods.
Machine Learning in Practice Lecture 10 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
CS 2750: Machine Learning The Bias-Variance Tradeoff Prof. Adriana Kovashka University of Pittsburgh January 13, 2016.
Roberto Battiti, Mauro Brunato
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Computational Intelligence: Methods and Applications Lecture 14 Bias-variance tradeoff – model selection. Włodzisław Duch Dept. of Informatics, UMK Google:
Data Science Credibility: Evaluating What’s Been Learned
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Roberto Battiti, Mauro Brunato
Machine Learning – Classification David Fenyő
Chapter 7. Classification and Prediction
CEE 6410 Water Resources Systems Analysis
Evaluating Classifiers
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Line Fitting James Hayes.
Boosting and Additive Trees
Classification with Perceptrons Reading:
Dipartimento di Ingegneria «Enzo Ferrari»,
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
A special case of calibration
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Machine Learning Techniques for Data Mining
Roberto Battiti, Mauro Brunato
The basic notions related to machine learning
CS 2750: Machine Learning Line Fitting + Bias-Variance Trade-off
Roberto Battiti, Mauro Brunato
Large Scale Support Vector Machines
Learning Algorithm Evaluation
ML – Lecture 3B Deep NN.
Model Evaluation and Selection
Overfitting and Underfitting
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
Model generalization Brief summary of methods
Machine learning overview
Evaluating Classifiers
The Bias-Variance Trade-Off
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
INTRODUCTION TO Machine Learning 3rd Edition
Introduction to Machine learning
COSC 4368 Intro Supervised Learning Organization
Support Vector Machines 2
CS249: Neural Language Model
Evaluation David Kauchak CS 158 – Fall 2019.
Presentation transcript:

Roberto Battiti, Mauro Brunato Roberto Battiti, Mauro Brunato. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligent-optimization.org/LIONbook © Roberto Battiti and Mauro Brunato , 2015, all rights reserved. Slides can be used and modified for classroom usage, provided that the attribution (link to book website) is kept.

Chap.3 Learning requires a method Data Mining, noun 1. Torturing the data until it confesses. . . and if you torture it long enough, you can get it to confess to anything.

What is learning? Unifying different cases by discovering the underlying explanatory laws. Learning from examples is only a means to reach the real goal: generalization, the capability of explaining new cases

Supervised learning architecture: feature extraction and classification

Performance estimation If the goal is generalization, estimating the performance has to be done with extreme care Feature extraction  (xi; yi), i = 1;...; L Classification Regression Output can be probability

Learning from labeled examples: minimization and generalization A flexible model f(x;w), where the flexibility is given by some tunable parameters (or weights) w determination of the best parameters is fully automated, this is why the method is called machine learning after all

Flexible model with internal parameters (mental images)

Learning from labeled examples: minimization and generalization (2) fix the free parameters by demanding that the learned model works correctly on the examples in the training set. power of optimization: we start by defining an error measure to be minimized, an automated optimization process to determine optimal parameters

Learning from labeled examples: minimization and generalization (3) suitable error measure is the sum of the errors between the correct answer (given by the example label) and the outcome predicted if the function is smooth one can discover points of low altitude by being blindfolded and parachuted to a random initial point… (gradient descent)

RMS (root mean square) error function Indivisual errors Square Average (Sum and divide) Square root is optional... (optimizing sum of squares or its square root leads to the sam eresult)

Bias-Variance dilemma minimization of an error function is a first critical component, but not the only one. We are interested in generalization  Model complexity matters!

Bias-Variance dilemma Models with too few parameters are inaccurate because of a large bias: they lack flexibility. Models with too many parameters are inaccurate because of a large variance: they are too sensitive to the sample details (changes in the details will produce huge variations). Identifying the best model requires identifying the proper “model complexity”, i.e., the proper architecture and number of parameters.

Learn, validate, test! careful experimental procedures to measure the effectiveness of the learning process. It is a terrible mistake to measure the performance of the learning systems on the same examples used for training The test set is used only once for a final measure of performance.

Learn, validate, test!

K-fold Cross-validation the original sample is randomly partitioned into K subsamples. A single subsample is used as the validation data for testing, and the remaining K - 1 subsamples are used as training data. The cross-validation process is then repeated K times (the folds), with each of the K subsamples used exactly once as the validation data. Validation results are averaged

K-fold Cross-validation Advanced: stratification helps in classification (subdivide each class)

Errors of different kinds

Errors of different kinds Accuracy is defined as the fraction of correct answers over the total Precision as the fraction of correct answers over the number of retrieved (positive) cases Recall is computed as the fraction of correct answers over the number of relevant (true positive) cases.

Gist The goal of machine learning is to use a set of training examples to realize a system which will correctly generalize to new cases, in the same context but not seen during learning. ML learns, i.e., determines appropriate values for the free parameters of a flexible model, by automatically minimizing a measure of the error on the example set, possibly corrected to discourage complex models, and therefore improving the chances of correct generalization. The output value of the system can be a class (classification), or a number (regression). In some cases having as output the probability for a class increases flexibility of usage.

Gist Accurate classifiers can be built without any knowledge elicitation phase, just starting from an abundant and representative set of example data. This is a dramatic paradigm change. ML is very powerful but requires a strict method (a kind of “pedagogy” of ML). For sure, never estimate performance on the training set — this is a mortal sin: be aware that re-using validation data will create optimistic estimates. If examples are scarce, use cross-validation to show off that you are an expert ML user. To be on the safe side, set away some test examples and use them only once at the end to estimate performance. There is no single way to measure the performance of a model, different kinds of mistakes can have very different costs. Accuracy, precision and recall are some possibilities, a confusion matrix is giving the complete picture for more classes.