R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.

Slides:



Advertisements
Similar presentations
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Advertisements

Learning Algorithm Evaluation
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Model generalization Test error Bias, variance and complexity
Indian Statistical Institute Kolkata
The loss function, the normal equation,
Evaluation.
Credibility: Evaluating what’s been learned. Evaluation: the key to success How predictive is the model we learned? Error on the training data is not.
Supervised classification performance (prediction) assessment Dr. Huiru Zheng Dr. Franscisco Azuaje School of Computing and Mathematics Faculty of Engineering.
Evaluation.
Experimental Evaluation
General Mining Issues a.j.m.m. (ton) weijters Overfitting Noise and Overfitting Quality of mined models (some figures are based on the ML-introduction.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Today Evaluation Measures Accuracy Significance Testing
Model Assessment and Selection Florian Markowetz & Rainer Spang Courses in Practical DNA Microarray Analysis.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
Evaluating Classifiers
Sampling : Error and bias. Sampling definitions  Sampling universe  Sampling frame  Sampling unit  Basic sampling unit or elementary unit  Sampling.
CLassification TESTING Testing classifier accuracy
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 5 of Data Mining by I. H. Witten, E. Frank and M. A. Hall 報告人:黃子齊
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
Outline 1-D regression Least-squares Regression Non-iterative Least-squares Regression Basis Functions Overfitting Validation 2.
Last lecture summary. Basic terminology tasks – classification – regression learner, algorithm – each has one or several parameters influencing its behavior.
Experimental Evaluation of Learning Algorithms Part 1.
Learning from Observations Chapter 18 Through
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 5.
CpSc 810: Machine Learning Evaluation of Classifier.
Manu Chandran. Outline Background and motivation Over view of techniques Cross validation Bootstrap method Setting up the problem Comparing AIC,BIC,Crossvalidation,Bootstrap.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
The Bias-Variance Trade-Off Oliver Schulte Machine Learning 726.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
WEKA Machine Learning Toolbox. You can install Weka on your computer from
Weka Just do it Free and Open Source ML Suite Ian Witten & Eibe Frank University of Waikato New Zealand.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
CSCI 347, Data Mining Evaluation: Cross Validation, Holdout, Leave-One-Out Cross Validation and Bootstrapping, Sections 5.3 & 5.4, pages
Classification Ensemble Methods 1
Validation methods.
Chapter 5: Credibility. Introduction Performance on the training set is not a good indicator of performance on an independent set. We need to predict.
CS 2750: Machine Learning The Bias-Variance Tradeoff Prof. Adriana Kovashka University of Pittsburgh January 13, 2016.
Roberto Battiti, Mauro Brunato
Machine Learning in Practice Lecture 2 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Computational Intelligence: Methods and Applications Lecture 15 Model selection and tradeoffs. Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Evaluating Classifiers. Reading for this topic: T. Fawcett, An introduction to ROC analysis, Sections 1-4, 7 (linked from class website)
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 5: Credibility: Evaluating What’s Been Learned.
Computacion Inteligente Least-Square Methods for System Identification.
Pattern recognition – basic concepts. Sample input attribute, attribute, feature, input variable, independent variable (atribut, rys, příznak, vstupní.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Data Science Credibility: Evaluating What’s Been Learned
Evaluating Classifiers
Boosting and Additive Trees
Classification with Perceptrons Reading:
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
CS 2750: Machine Learning Line Fitting + Bias-Variance Trade-off
Roberto Battiti, Mauro Brunato
Learning Algorithm Evaluation
Model Evaluation and Selection
Machine learning overview
Evaluating Classifiers
Presentation transcript:

R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb optimization.org/LIONbook © Roberto Battiti and Mauro Brunato, 2014, all rights reserved. Can be used and modified for classroom usage, provided that the attribution (link to book website) is kept.

Chap.3 Learning requires a method Data Mining, noun 1. Torturing the data until it confesses... and if you torture it long enough, you can get it to confess to anything.

What is learning? Unifying different cases by discovering the underlying explanatory laws. Learning from examples is only a means to reach the real goal: generalization, the capability of explaining new cases

Supervised learning architecture: feature extraction and classification

Performance estimation If the goal is generalization, estimating the performance has to be done with extreme care Feature extraction  (xi; yi), i = 1;...; L Classification Regression Output can be probability

Learning from labeled examples: minimization and generalization A flexible model f(x;w), where the flexibility is given by some tunable parameters (or weights) w determination of the best parameters is fully automated, this is why the method is called machine learning after all

Flexible model with internal parameters (mental images)

Learning from labeled examples: minimization and generalization (2) fix the free parameters by demanding that the learned model works correctly on the examples in the training set. power of optimization: we start by defining an error measure to be minimized, an automated optimization process to determine optimal parameters

Learning from labeled examples: minimization and generalization (3) suitable error measure is the sum of the errors between the correct answer (given by the example label) and the outcome predicted if the function is smooth one can discover points of low altitude by being blindfolded and parachuted to a random initial point… (gradient descent)

RMS (root mean square) error function Indivisual errors Square Average (Sum and divide) Square root is optional... (optimizing sum of squares or its square root leads to the sam eresult)

Bias-Variance dilemma minimization of an error function is a first critical component, but not the only one. We are interested in generalization  Model complexity matters!

Bias-Variance dilemma 1.Models with too few parameters are inaccurate because of a large bias: they lack flexibility. 2.Models with too many parameters are inaccurate because of a large variance: they are too sensitive to the sample details (changes in the details will produce huge variations). 3.Identifying the best model requires identifying the proper “model complexity”, i.e., the proper architecture and number of parameters.

Learn, validate, test! careful experimental procedures to measure the effectiveness of the learning process. It is a terrible mistake to measure the performance of the learning systems on the same examples used for training The test set is used only once for a final measure of performance.

Learn, validate, test!

K-fold Cross-validation the original sample is randomly partitioned into K subsamples. A single subsample is used as the validation data for testing, and the remaining K - 1 subsamples are used as training data. The cross-validation process is then repeated K times (the folds), with each of the K subsamples used exactly once as the validation data. Validation results are averaged

K-fold Cross-validation Advanced: stratification helps in classification (subdivide each class)

Errors of different kinds

Errors of different kinds Accuracy is defined as the fraction of correct answers over the total Precision as the fraction of correct answers over the number of retrieved (positive) cases Recall is computed as the fraction of correct answers over the number of relevant (true positive) cases.

Gist The goal of machine learning is to use a set of training examples to realize a system which will correctly generalize to new cases, in the same context but not seen during learning. ML learns, i.e., determines appropriate values for the free parameters of a flexible model, by automatically minimizing a measure of the error on the example set, possibly corrected to discourage complex models, and therefore improving the chances of correct generalization. The output value of the system can be a class (classification), or a number (regression). In some cases having as output the probability for a class increases flexibility of usage.

Gist Accurate classifiers can be built without any knowledge elicitation phase, just starting from an abundant and representative set of example data. This is a dramatic paradigm change. ML is very powerful but requires a strict method (a kind of “pedagogy” of ML). For sure, never estimate performance on the training set — this is a mortal sin: be aware that re-using validation data will create optimistic estimates. If examples are scarce, use cross-validation to show off that you are an expert ML user. To be on the safe side, set away some test examples and use them only once at the end to estimate performance. There is no single way to measure the performance of a model, different kinds of mistakes can have very different costs. Accuracy, precision and recall are some possibilities, a confusion matrix is giving the complete picture for more classes.