CS 478 - Instance Based Learning1 Instance Based Learning.

Slides:



Advertisements
Similar presentations
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
Advertisements

Data Mining Classification: Alternative Techniques
Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Support Vector Machines
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
K nearest neighbor and Rocchio algorithm
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Instance Based Learning
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
Nearest Neighbor Classifiers other names: –instance-based learning –case-based learning (CBL) –non-parametric learning –model-free learning.
Radial Basis Function (RBF) Networks
Radial-Basis Function Networks
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks II PROF. DR. YUSUF OYSAL.
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Module 04: Algorithms Topic 07: Instance-Based Learning
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Artificial Neural Networks Shreekanth Mandayam Robi Polikar …… …... … net k   
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
Basic Data Mining Technique
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
1 Instance Based Learning Ata Kaban The University of Birmingham.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Fast Learning in Networks of Locally-Tuned Processing Units John Moody and Christian J. Darken Yale Computer Science Neural Computation 1, (1989)
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
CS Machine Learning Instance Based Learning (Adapted from various sources)
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Machine Learning Supervised Learning Classification and Regression
Instance Based Learning
K Nearest Neighbors and Instance-based methods
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Nearest-Neighbor Classifiers
Neuro-Computing Lecture 4 Radial Basis Function Network
Instance Based Learning
Chap 8. Instance Based Learning
Chapter 8: Generalization and Function Approximation
Machine Learning: UNIT-4 CHAPTER-1
Introduction to Radial Basis Function Networks
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

CS Instance Based Learning1 Instance Based Learning

CS Instance Based Learning2 Instance Based Learning Classify based on local similarity Ranges from simple nearest neighbor to case-based and analogical reasoning Use local information near the current query instance to decide the classification of that instance As such can represent quite complex decision surfaces in a simple manner

k-Nearest Neighbor Approach Simply store all (or some representative subset) of the examples in the training set. When desiring to generalize on a new instance, measure the distance from the new instance to one or more stored instances which vote to decide the class of the new instance. No need to pre-process a specific hypothesis (Lazy vs. Eager learning) – Fast learning – Can be slow during execution and require significant storage – Some models index the data or reduce the instances stored

CS Instance Based Learning4 k-Nearest Neighbor (cont.) Naturally supports real valued attributes Typically use Euclidean distance Nominal/unknown attributes can just be a 1/0 distance (more on other distance metrics later) The output class for the query instance is set to the most common class of its k nearest neighbors where  (x,y) = 1 if x = y, else 0 k greater than 1 is more noise resistant, but a very large k would lead to less accuracy as less relevant neighbors have more influence (common values: k=3, k=5)

CS Instance Based Learning5 k-Nearest Neighbor (cont.) Can also do distance weighted voting where the strength of a neighbors influence is proportional to its distance Inverse of distance squared is a common weight Gaussian is another common distance weight In this case can let k be larger (even all points if desired), because the more distant points have negligible influence

CS Instance Based Learning6 Regression with k-nn Can also do regression by letting the output be the weighted mean of the k nearest neighbors For distance weighted regression Where f(x) is the output value for instance x

CS Instance Based Learning7 Attribute Weighting One of the main weaknesses of nearest neighbor is irrelevant features, since they can dominate the distance Can create algorithms which weight the attributes (Note that backprop and ID3 etc, do higher order weighting of features) No longer lazy evaluation since you need to come up with a portion of your hypothesis (attribute weights) before generalizing Still an open area of research – Should weighting be local or global? – What is the best method, etc.?

CS Instance Based Learning8 Reduction Techniques Wilson, D. R. and Martinez, T. R., Reduction Techniques for Exemplar-Based Learning Algorithms, Machine Learning Journal, vol. 38, no. 3, pp , Create a subset or other representative set of prototype nodes Approaches – Leave-one-out reduction - Drop instance if it would still be classified correctly – Growth algorithm - Only add instance if it is not already classified correctly - both order dependent – More global optimizing approaches – Just keep central points – Just keep border points (pre-process noisy instances) – Drop 5 (Wilson & Martinez) maintains almost full accuracy with approximately 15% of the original instances

CS Instance Based Learning9

10 Radial Basis Function (RBF) Networks Each prototype node computes a distance based kernel function (Gaussian is common) Prototype nodes form a hidden layer in a neural network Train top layer with simple delta rule to get outputs Thus, prototype nodes learn weightings for each class 2 A 3 xy 1 B 4 1 x y 2 3 4

CS Instance Based Learning11 Radial Basis Function Networks Number of nodes and placement (means) Sphere of influence (deviation) – Too small - no generalization, should have some overlap – Too large - saturation, lose local effects, long training Output layer weights - linear or non-linear nodes – Delta rule variations – Direct matrix weight calculation

CS Instance Based Learning12 Node Placement One for each instance of the training set Random subset of instances Clustering - Unsupervised or supervised - k-means style vs. constructive Genetic Algorithms Random Coverage - Curse of Dimensionality Node adjustment - Competitive Learning style Dynamic addition and deletion of nodes

CS Instance Based Learning13 RBF vs. BP Line vs. Sphere - mix-and-match approaches Potential Faster Training - nearest neighbor localization - Yet more data and hidden nodes typically needed Local vs Global, less extrapolation (ala BP), have reject capability (avoid false positives) RBF can have problems with irrelevant features just like nearest neighbor

CS Instance Based Learning14 Distance Metrics Wilson, D. R. and Martinez, T. R., Improved Heterogeneous Distance Functions, Journal of Artificial Intelligence Research, vol. 6, no. 1, pp. 1-34, Normalization of features Main question: How best to handle nominal inputs

CS Instance Based Learning15

CS Instance Based Learning16

CS Instance Based Learning17

CS Instance Based Learning18

CS Instance Based Learning19

CS Instance Based Learning20

CS Instance Based Learning21 Instance Based Learning Assignment See html html