Instance Based Learning

Slides:



Advertisements
Similar presentations
1 Classification using instance-based learning. 3 March, 2000Advanced Knowledge Management2 Introduction (lazy vs. eager learning) Notion of similarity.
Advertisements

Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Support Vector Machines
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Instance Based Learning
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
Statistical Classification Rong Jin. Classification Problems X Input Y Output ? Given input X={x 1, x 2, …, x m } Predict the class label y  Y Y = {-1,1},
CMPUT 466/551 Principal Source: CMU
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
Instance Based Learning
Università di Milano-Bicocca Laurea Magistrale in Informatica Corso di APPRENDIMENTO E APPROSSIMAZIONE Lezione 8 - Instance based learning Prof. Giancarlo.
K nearest neighbor and Rocchio algorithm
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Instance based learning K-Nearest Neighbor Locally weighted regression Radial basis functions.
Learning from Observations Chapter 18 Section 1 – 4.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Instance-Based Learning
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor.
Nearest Neighbor Classifiers other names: –instance-based learning –case-based learning (CBL) –non-parametric learning –model-free learning.
CS Instance Based Learning1 Instance Based Learning.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Radial Basis Function Networks
Module 04: Algorithms Topic 07: Instance-Based Learning
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
1 Instance Based Learning Ata Kaban The University of Birmingham.
CpSc 881: Machine Learning Instance Based Learning.
CpSc 810: Machine Learning Instance Based Learning.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, November 23, 1999.
KNN Classifier.  Handed an instance you wish to classify  Look around the nearby region to see what other classes are around  Whichever is most common—make.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Thursday, 05 April 2007 William.
Meta-learning for Algorithm Recommendation Meta-learning for Algorithm Recommendation Background on Local Learning Background on Algorithm Assessment Algorithm.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 15: Mixtures of Experts Geoffrey Hinton.
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
CS 8751 ML & KDDInstance Based Learning1 k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning.
1 Instance Based Learning Soongsil University Intelligent Systems Lab.
Instance Based Learning
Data Mining: Concepts and Techniques (3rd ed
Data Mining Lecture 11.
K Nearest Neighbor Classification
یادگیری بر پایه نمونه Instance Based Learning Instructor : Saeed Shiry
Instance Based Learning
Chap 8. Instance Based Learning
Chapter 8: Generalization and Function Approximation
Machine Learning: UNIT-4 CHAPTER-1
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Presentation transcript:

Instance Based Learning k-Nearest Neighbor Locally weighted regression Radial basis functions Case-based reasoning Lazy and eager learning CS 5751 Machine Learning Chapter 8 Instance Based Learning

Instance-Based Learning CS 5751 Machine Learning Chapter 8 Instance Based Learning

When to Consider Nearest Neighbor Instance map to points in Rn Less than 20 attributes per instance Lots of training data Advantages Training is very fast Learn complex target functions Do not lose information Disadvantages Slow at query time Easily fooled by irrelevant attributes CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning k-NN Classification CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning Behavior in the Limit Define p(x) as probability that instance x will be labeled 1 (positive) versus 0 (negative) Nearest Neighbor As number of training examples approaches infinity, approaches Gibbs Algorithm Gibbs: with probability p(x) predict 1, else 0 k-Nearest Neighbor: As number of training examples approaches infinity and k gets large, approaches Bayes optimal Bayes optimal: if p(x) > 0.5 then predict 1, else 0 Note Gibbs has at most twice the expected error of Bayes optimal CS 5751 Machine Learning Chapter 8 Instance Based Learning

Distance-Weighted k-NN CS 5751 Machine Learning Chapter 8 Instance Based Learning

Curse of Dimensionality Imagine instances described by 20 attributes, but only 2 are relevant to target function Curse of dimensionality: nearest neighbor is easily misled when high-dimensional X One approach: Stretch jth axis by weight zj, where z1,z2,…,zn chosen to minimize prediction error Use cross-validation to automatically choose weights z1,z2,…,zn Note setting zj to zero eliminates dimension j altogether see (Moore and Lee, 1994) CS 5751 Machine Learning Chapter 8 Instance Based Learning

Locally Weighted Regression CS 5751 Machine Learning Chapter 8 Instance Based Learning

Radial Basis Function Networks Global approximation to target function, in terms of linear combination of local approximations Used, for example, in image classification A different kind of neural network Closely related to distance-weighted regression, but “eager” instead of “lazy” CS 5751 Machine Learning Chapter 8 Instance Based Learning

Radial Basis Function Networks CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning Training RBF Networks Q1: What xu to use for kernel function Ku(d(xu,x))? Scatter uniformly through instance space Or use training instances (reflects instance distribution) Q2: How to train weights (assume here Gaussian Ku)? First choose variance (and perhaps mean) for each Ku e.g., use EM Then hold Ku fixed, and train linear output layer efficient methods to fit linear function CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning Case-Based Reasoning Can apply instance-based learning even when X| Rn  need different “distance” metric Case-Based Reasoning is instance-based learning applied to instances with symbolic logic descriptions: ((user-complaint error53-on-shutdown) (cpu-model PowerPC) (operating-system Windows) (network-connection PCIA) (memory 48meg) (installed-applications Excel Netscape VirusScan) (disk 1Gig) (likely-cause ???)) CS 5751 Machine Learning Chapter 8 Instance Based Learning

Case-Based Reasoning in CADET CADET: 75 stored examples of mechanical devices each training example: <qualitative function, mechanical structure> new query: desired function target value: mechanical structure for this function Distance metric: match qualitative function descriptions CS 5751 Machine Learning Chapter 8 Instance Based Learning

Case-Based Reasoning in CADET CS 5751 Machine Learning Chapter 8 Instance Based Learning

Case-Based Reasoning in CADET Instances represented by rich structural descriptions Multiple cases retrieved (and combined) to form solution to new problem Tight coupling between case retrieval and problem solving Bottom line: Simple matching of cases useful for tasks such as answering help-desk queries Area of ongoing research CS 5751 Machine Learning Chapter 8 Instance Based Learning

Lazy and Eager Learning Lazy: wait for query before generalizing k-Nearest Neighbor, Case-Based Reasoning Eager: generalize before seeing query Radial basis function networks, ID3, Backpropagation, etc. Does it matter? Eager learner must create global approximation Lazy learner can create many local approximations If they use same H, lazy can represent more complex functions (e.g., consider H=linear functions) CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning kd-trees (Moore) Eager version of k-Nearest Neighbor Idea: decrease time to find neighbors train by constructing a lookup (kd) tree recursively subdivide space ignore class of points lots of possible mechanisms: grid, maximum variance, etc. when looking for nearest neighbor search tree nearest neighbor can be found in log(n) steps k nearest neighbors can be found by generalizing process (still in log(n) steps if k is constant) Slower training but faster classification CS 5751 Machine Learning Chapter 8 Instance Based Learning

Chapter 8 Instance Based Learning kd Tree CS 5751 Machine Learning Chapter 8 Instance Based Learning

Instance Based Learning Summary Lazy versus Eager learning lazy - work done at testing time eager -work done at training time instance based sometimes lazy k-Nearest Neighbor (k-nn) lazy classify based on k nearest neighbors key: determining neighbors variations: distance weighted combination locally weighted regression limitation: curse of dimensionality “stretching” dimensions CS 5751 Machine Learning Chapter 8 Instance Based Learning

Instance Based Learning Summary k-d trees (eager version of k-nn) structure built at train time to quickly find neighbors Radial Basis Function (RBF) networks (eager) units active in region (sphere) of space key: picking/training kernel functions Case-Based Reasoning (CBR) generally lazy nearest neighbor when no continuos features may have other types of features: structural (graphs in CADET) CS 5751 Machine Learning Chapter 8 Instance Based Learning