CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.

Slides:



Advertisements
Similar presentations
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
Advertisements

Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e © The MIT Press (V1.0) ETHEM ALPAYDIN © The MIT Press, 2010
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Statistical Classification Rong Jin. Classification Problems X Input Y Output ? Given input X={x 1, x 2, …, x m } Predict the class label y  Y Y = {-1,1},
CMPUT 466/551 Principal Source: CMU
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
Università di Milano-Bicocca Laurea Magistrale in Informatica Corso di APPRENDIMENTO E APPROSSIMAZIONE Lezione 8 - Instance based learning Prof. Giancarlo.
Kernel methods - overview
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
Learning from Observations Chapter 18 Section 1 – 4.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance Based Learning
Instance-Based Learning
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
Nearest Neighbor Classifiers other names: –instance-based learning –case-based learning (CBL) –non-parametric learning –model-free learning.
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Radial Basis Function Networks
8/10/ RBF NetworksM.W. Mak Radial Basis Function Networks 1. Introduction 2. Finding RBF Parameters 3. Decision Surface of RBF Networks 4. Comparison.
Module 04: Algorithms Topic 07: Instance-Based Learning
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
1 Instance Based Learning Ata Kaban The University of Birmingham.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, November 23, 1999.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 05 April 2007 William.
Meta-learning for Algorithm Recommendation Meta-learning for Algorithm Recommendation Background on Local Learning Background on Algorithm Assessment Algorithm.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
General-Purpose Learning Machine
Instance Based Learning
Data Mining: Concepts and Techniques (3rd ed
Data Mining Lecture 11.
K Nearest Neighbor Classification
یادگیری بر پایه نمونه Instance Based Learning Instructor : Saeed Shiry
Instance Based Learning
Chap 8. Instance Based Learning
Chapter 8: Generalization and Function Approximation
Machine Learning: UNIT-4 CHAPTER-1
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning

CS 8751 ML & KDDInstance Based Learning2 Instance-Based Learning

CS 8751 ML & KDDInstance Based Learning3 When to Consider Nearest Neighbor Instance map to points in R n Less than 20 attributes per instance Lots of training data Advantages Training is very fast Learn complex target functions Do not lose information Disadvantages Slow at query time Easily fooled by irrelevant attributes

CS 8751 ML & KDDInstance Based Learning4 k-NN Classification

CS 8751 ML & KDDInstance Based Learning5 Behavior in the Limit Define p(x) as probability that instance x will be labeled 1 (positive) versus 0 (negative) Nearest Neighbor As number of training examples approaches infinity, approaches Gibbs Algorithm Gibbs: with probability p(x) predict 1, else 0 k-Nearest Neighbor: As number of training examples approaches infinity and k gets large, approaches Bayes optimal Bayes optimal: if p(x) > 0.5 then predict 1, else 0 Note Gibbs has at most twice the expected error of Bayes optimal

CS 8751 ML & KDDInstance Based Learning6 Distance-Weighted k-NN

CS 8751 ML & KDDInstance Based Learning7 Curse of Dimensionality Imagine instances described by 20 attributes, but only 2 are relevant to target function Curse of dimensionality: nearest neighbor is easily misled when high-dimensional X One approach: Stretch jth axis by weight z j, where z 1,z 2,…,z n chosen to minimize prediction error Use cross-validation to automatically choose weights z 1,z 2,…,z n Note setting z j to zero eliminates dimension j altogether see (Moore and Lee, 1994)

CS 8751 ML & KDDInstance Based Learning8 Locally Weighted Regression

CS 8751 ML & KDDInstance Based Learning9 Radial Basis Function Networks Global approximation to target function, in terms of linear combination of local approximations Used, for example, in image classification A different kind of neural network Closely related to distance-weighted regression, but “eager” instead of “lazy”

CS 8751 ML & KDDInstance Based Learning10 Radial Basis Function Networks

CS 8751 ML & KDDInstance Based Learning11 Training RBF Networks Q1: What x u to use for kernel function K u (d(x u,x))? Scatter uniformly through instance space Or use training instances (reflects instance distribution) Q2: How to train weights (assume here Gaussian K u )? First choose variance (and perhaps mean) for each K u –e.g., use EM Then hold K u fixed, and train linear output layer –efficient methods to fit linear function

CS 8751 ML & KDDInstance Based Learning12 Case-Based Reasoning Can apply instance-based learning even when X | R n  need different “distance” metric Case-Based Reasoning is instance-based learning applied to instances with symbolic logic descriptions: ((user-complaint error53-on-shutdown) (cpu-model PowerPC) (operating-system Windows) (network-connection PCIA) (memory 48meg) (installed-applications Excel Netscape VirusScan) (disk 1Gig) (likely-cause ???))

CS 8751 ML & KDDInstance Based Learning13 Case-Based Reasoning in CADET CADET: 75 stored examples of mechanical devices each training example: new query: desired function target value: mechanical structure for this function Distance metric: match qualitative function descriptions

CS 8751 ML & KDDInstance Based Learning14 Case-Based Reasoning in CADET

CS 8751 ML & KDDInstance Based Learning15 Case-Based Reasoning in CADET Instances represented by rich structural descriptions Multiple cases retrieved (and combined) to form solution to new problem Tight coupling between case retrieval and problem solving Bottom line: Simple matching of cases useful for tasks such as answering help-desk queries Area of ongoing research

CS 8751 ML & KDDInstance Based Learning16 Lazy and Eager Learning Lazy: wait for query before generalizing k-Nearest Neighbor, Case-Based Reasoning Eager: generalize before seeing query Radial basis function networks, ID3, Backpropagation, etc. Does it matter? Eager learner must create global approximation Lazy learner can create many local approximations If they use same H, lazy can represent more complex functions (e.g., consider H=linear functions)

CS 8751 ML & KDDInstance Based Learning17 kd-trees (Moore) Eager version of k-Nearest Neighbor Idea: decrease time to find neighbors –train by constructing a lookup (kd) tree –recursively subdivide space ignore class of points lots of possible mechanisms: grid, maximum variance, etc. –when looking for nearest neighbor search tree –nearest neighbor can be found in log(n) steps –k nearest neighbors can be found by generalizing process (still in log(n) steps if k is constant) Slower training but faster classification

CS 8751 ML & KDDInstance Based Learning18 kd Tree