Roberto Battiti, Mauro Brunato

Slides:



Advertisements
Similar presentations
Nonparametric Methods: Nearest Neighbors
Advertisements

Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Data Mining Classification: Alternative Techniques
Support Vector Machines and Margins
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Classification and Decision Boundaries
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
Instance Based Learning IB1 and IBK Small section in chapter 20.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
INSTANCE-BASE LEARNING
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Linear Discrimination Reading: Chapter 2 of textbook.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
Roberto Battiti, Mauro Brunato
CS Machine Learning Instance Based Learning (Adapted from various sources)
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
CHAPTER 8: Nonparametric Methods Alpaydin transparencies significantly modified, extended and changed by Ch. Eick Last updated: March 4, 2011.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Unveiling Zeus Automated Classification of Malware Samples Abedelaziz Mohaisen Omar Alrawi Verisign Inc, VA, USA Verisign Labs, VA, USA
Roberto Battiti, Mauro Brunato
Who am I? Work in Probabilistic Machine Learning Like to teach 
Roberto Battiti, Mauro Brunato
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Instance Based Learning
Ch8: Nonparametric Methods
Classification Nearest Neighbor
The Elements of Statistical Learning
Machine Learning Basics
ECE 5424: Introduction to Machine Learning
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Instance Based Learning (Adapted from various sources)
Machine Learning Week 1.
Roberto Battiti, Mauro Brunato
K Nearest Neighbor Classification
Roberto Battiti, Mauro Brunato
Classification Nearest Neighbor
Roberto Battiti, Mauro Brunato
Nearest-Neighbor Classifiers
Roberto Battiti, Mauro Brunato
Instance Based Learning
COSC 4335: Other Classification Techniques
Chap 8. Instance Based Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Machine Learning: UNIT-4 CHAPTER-1
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Data Mining CSCI 307, Spring 2019 Lecture 11
Presentation transcript:

Roberto Battiti, Mauro Brunato Roberto Battiti, Mauro Brunato. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligent-optimization.org/LIONbook © Roberto Battiti and Mauro Brunato , 2015, all rights reserved. Slides can be used and modified for classroom usage, provided that the attribution (link to book website) is kept.

Chap.2 Lazy learning: nearest neighbors Natura non facit saltus Nature does not make jumps

Learning from examples written characters (“a”,“b”,“c”. . . ) labeled examples Memorization is different from extracting the underlying patterns and regularities Generalization is the key

Learning from examples Mushroom hunting requires classifying edible and poisonous species.

Learning from examples In supervised learning a system is trained by a supervisor (teacher) giving labeled examples. Each example is an array, a vector of input parameters x called features with an associated output label y.

Lazy “nearest neighbor” method of machine learning. Lazy beginners in mushroom picking

Nearest-Neighbors Methods nearest-neighbors basic form of learning, also related to instance-based learning, case-based or memory-based, pattern patching Classify a new example according to the nearest neighbor (considering input values)

Neighbor  Neigbors k-nearest-neighbors (KNN) Majority or unanimity rule

Regression regression (the prediction of a real number, like the content of poison in a mushroom), the output can be obtained as a simple average

Weighted K-nearest neighbors (WKNN) Closer neighbors receive a higher weight (and more influence to derive the output value)

Kernels and locally weighted regression Kernel methods and locally-weighted regression can be seen as flexible and smooth generalizations of the nearest-neighbors idea Instead of applying a brute exclusion of the distant points, all points contribute to the output but with a significance (“weight”) related to their distance from the query point.

Gist KNN (K Nearest Neighbors) is a primitive and lazy form of machine learning: just store all training examples into memory. When a new input to be evaluated appears, search in memory for the K closest examples stored. Read their output and derive the output for the new input by majority or averaging. Laziness during training causes long response times when searching a very large memory storage. KNN works in many real-world cases because similar inputs are usually related to similar outputs, a basic hypothesis in machine learning. It is similar to some “case-based” human reasoning processes. Although simple and brutal, it can be surprisingly effective in many cases.