R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.

Slides:



Advertisements
Similar presentations
Nonparametric Methods: Nearest Neighbors
Advertisements

Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Support Vector Machines and Margins
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance Based Learning
Data Mining Classification: Alternative Techniques
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
CES 514 – Data Mining Lec 9 April 14 Mid-term k nearest neighbor.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
Roberto Battiti, Mauro Brunato
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb 2014.
K Nearest Neighborhood (KNNs)
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
K Nearest Neighbors Saed Sayad 1www.ismartsoft.com.
ECE 5984: Introduction to Machine Learning Dhruv Batra Virginia Tech Topics: –(Finish) Nearest Neighbour Readings: Barber 14 (kNN)
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Predictive Application- Performance Modeling in a Computational Grid Environment (HPDC ‘99) Nirav Kapadia, José Fortes, Carla Brodley ECE, Purdue Presented.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
6.S093 Visual Recognition through Machine Learning Competition Image by kirkh.deviantart.com Joseph Lim and Aditya Khosla Acknowledgment: Many slides from.
Roberto Battiti, Mauro Brunato
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
CHAPTER 8: Nonparametric Methods Alpaydin transparencies significantly modified, extended and changed by Ch. Eick Last updated: March 4, 2011.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
Outline Time series prediction Find k-nearest neighbors Lag selection Weighted LS-SVM.
Unveiling Zeus Automated Classification of Malware Samples Abedelaziz Mohaisen Omar Alrawi Verisign Inc, VA, USA Verisign Labs, VA, USA
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
Instance Based Learning (Adapted from various sources)
Roberto Battiti, Mauro Brunato
Roberto Battiti, Mauro Brunato
K Nearest Neighbor Classification
Roberto Battiti, Mauro Brunato
Classification Nearest Neighbor
Roberto Battiti, Mauro Brunato
Nearest-Neighbor Classifiers
Roberto Battiti, Mauro Brunato
Instance Based Learning
COSC 4335: Other Classification Techniques
Presentation transcript:

R OBERTO B ATTITI, M AURO B RUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Feb optimization.org/LIONbook © Roberto Battiti and Mauro Brunato, 2014, all rights reserved. Can be used and modified for classroom usage, provided that the attribution (link to book website) is kept.

Chap.2 Lazy learning: nearest neighbors Natura non facit saltus Nature does not make jumps

Learning from examples written characters (“a”,“b”,“c”... ) labeled examples Memorization is different from extracting the underlying patterns and regularities Generalization is the key

Learning from examples Mushroom hunting requires classifying edible and poisonous species.

Learning from examples In supervised learning a system is trained by a supervisor (teacher) giving labeled examples. Each example is an array, a vector of input parameters x called features with an associated output label y.

Lazy “nearest neighbor” method of machine learning. Lazy beginners in mushroom picking

Nearest-Neighbors Methods nearest-neighbors basic form of learning, also related to instance-based learning, case- based or memory-based, pattern patching Classify a new example according to the nearest neighbor (considering input values)

Neighbor  Neigbors k-nearest-neighbors (KNN) Majority or unanimity rule

Regression regression (the prediction of a real number, like the content of poison in a mushroom), the output can be obtained as a simple average

Weighted K-nearest neighbors (WKNN) Closer neighbors receive a higher weight (and more influence to derive the output value)

Kernels and locally weighted regression Kernel methods and locally-weighted regression can be seen as flexible and smooth generalizations of the nearest-neighbors idea Instead of applying a brute exclusion of the distant points, all points contribute to the output but with a significance (“weight”) related to their distance from the query point.

Gist KNN (K Nearest Neighbors) is a primitive and lazy form of machine learning: just store all training examples into memory. When a new input to be evaluated appears, search in memory for the K closest examples stored. Read their output and derive the output for the new input by majority or averaging. Laziness during training causes long response times when searching a very large memory storage. KNN works in many real-world cases because similar inputs are usually related to similar outputs, a basic hypothesis in machine learning. It is similar to some “case-based” human reasoning processes. Although simple and brutal, it can be surprisingly effective in many cases.