Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

DECISION TREES. Decision trees  One possible representation for hypotheses.
Machine Learning Instance Based Learning & Case Based Reasoning Exercise Solutions.
My name is Dustin Boswell and I will be presenting: Ensemble Methods in Machine Learning by Thomas G. Dietterich Oregon State University, Corvallis, Oregon.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
Instance Based Learning
Machine Learning Group University College Dublin Nearest Neighbour Classifiers Lazy v’s Eager k-NN Condensed NN.
Three kinds of learning
12 March 1999Dip HI KBS Module1 Machine Learning Lucy Hederman.
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
CES 514 – Data Mining Lec 9 April 14 Mid-term k nearest neighbor.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Nearest Neighbour Condensing and Editing David Claus February 27, 2004 Computer Vision Reading Group Oxford.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
CS Instance Based Learning1 Instance Based Learning.
Machine learning Image source:
Machine learning Image source:
K Nearest Neighborhood (KNNs)
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
Nearest Neighbor (NN) Rule & k-Nearest Neighbor (k-NN) Rule Non-parametric : Can be used with arbitrary distributions, No need to assume that the form.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
CLASSIFICATION: Ensemble Methods
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
1 Instance Based Learning Ata Kaban The University of Birmingham.
On the Role of Dataset Complexity in Case-Based Reasoning Derek Bridge UCC Ireland (based on work done with Lisa Cummins)
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 6: Nearest and k-nearest Neighbor Classification.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Data Mining and Decision Support
Data Mining By: Johan Johansson. Mining Techniques Association Rules Association Rules Decision Trees Decision Trees Clustering Clustering Nearest Neighbor.
基于实例的学习 上海交通大学 摘自 T. Mitchell Machine learning Instance based learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Introduction to Machine Learning
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Classification Nearest Neighbor
K Nearest Neighbor Classification
Nearest-Neighbor Classifiers
Instance Based Learning
COSC 4335: Other Classification Techniques
Chap 8. Instance Based Learning
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Machine Learning: UNIT-4 CHAPTER-1
Nearest Neighbor Classifiers
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers

What is Lazy Learning? Compare ANNs and CBR or k-NN classifier –Artificial Neural Networks are eager learners training examples compiled into a model at training time not available at runtime –CBR or k-Nearest Neighbour are lazy little offline learning done work deferred to runtime Compare conventional use of lazy-eager in computer science

Outline Classification problems Classification techniques k-Nearest Neighbour –Condense training Set –Feature selection –Feature weighting Ensemble techniques in ML

Classification problems Exemplar characterised by a set of features; decide class to which exemplar belongs Compare regression problems Exemplar characterised by a set of features; decide value of continuous output (dependant) variable

Classifying apples and pears To what class does this belong?

Distance/Similarity Function For query q and training set X (described by features F) compute d(x,q) for each x  X, where and where Category of q decided by its k Nearest Neighbours

k-NN and Noise 1-NN easy to implement –susceptible to noise a misclassification every time a noisy pattern retrieved k-NN with k  3 will overcome this

e.g. Pregnancy prediction

e.g. MVT Machine Vision for inspection of PCBs –components present or absent –solder joints good or bad

Components present? Absent Present

Characterise image as a set of features

Classification techniques Artificial Neural Networks –also good for non linear regression –black box development tricky users do not know what is going on Decision Trees –built using induction (information theoretic analysis) k-Nearest Neighbour classifiers –keep training examples, find k nearest at run time

Dimension reduction in k-NN Not all features required –noisy features a hindrance Some examples redundant –retrieval time depends on no. of examples p features q best features n covering examples m examples

Condensed NN D set of training samples Find E where E  D; NN rule used with E should be as good as with D choose x  D randomly, D  D \ {x}, E  {x}, DO learning?  FALSE, FOR EACH x  D classify x by NN using E, if classification incorrect then E  E  {x}, D  D \ {x}, learning  TRUE, WHILE (learning?  FALSE)

Condensed NN 100 examples 2 categories Different CNN solutions

Improving Condensed NN Sort data based on distance to nearest unlike neighbour A B –identify exemplars near decision surface –in diagram B more useful than A Different outcomes depending on data order –that’s a bad thing in an algorithm

Condensed NN 100 examples 2 categories Different CNN solutions CNN using NUN

m Feature selection Irrelevant features are noise: –make classification harder Extra features add to computation cost p

Ensemble techniques For the user with more machine cycles than they know what to do with Outcome Combiner Classifiers Build several classifiers –different training subsets –different feature subsets Aggregate results –voting vote based on generalisation error

Conclusions Finding a covering set of training data –very good solutions exist Compare with results of Ensemble techniques