Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)

Slides:



Advertisements
Similar presentations
Lecture 3-4: Classification & Clustering
Advertisements

Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Distance and Similarity Measures
1 Machine Learning: Lecture 7 Instance-Based Learning (IBL) (Based on Chapter 8 of Mitchell T.., Machine Learning, 1997)
Lazy vs. Eager Learning Lazy vs. eager learning
1er. Escuela Red ProTIC - Tandil, de Abril, Instance-Based Learning 4.1 Introduction Instance-Based Learning: Local approximation to the.
Classification and Decision Boundaries
Data Mining Classification: Alternative Techniques
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class.
CES 514 – Data Mining Lecture 8 classification (contd…)
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
Data Mining Classification: Alternative Techniques
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
Lecture outline Classification Naïve Bayes classifier Nearest-neighbor classifier.
1 Nearest Neighbor Learning Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Intro AI.
CES 514 – Data Mining Lec 9 April 14 Mid-term k nearest neighbor.
Aprendizagem baseada em instâncias (K vizinhos mais próximos)
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
INSTANCE-BASE LEARNING
The UNIVERSITY of Kansas EECS 800 Research Seminar Mining Biological Data Instructor: Luke Huan Fall, 2006.
CS Instance Based Learning1 Instance Based Learning.
Distance and Similarity Measures
Module 04: Algorithms Topic 07: Instance-Based Learning
Nearest Neighbor Learning
Learning: Nearest Neighbor Artificial Intelligence CMSC January 31, 2002.
DATA MINING LECTURE 11 Classification Nearest Neighbor Classification Support Vector Machines Logistic Regression Naïve Bayes Classifier Supervised Learning.
K Nearest Neighborhood (KNNs)
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
1 Data Mining Lecture 5: KNN and Bayes Classifiers.
K nearest neighbor classification Presented by Vipin Kumar University of Minnesota Based on discussion in "Intro to Data Mining" by Tan,
 2003, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci 9 Instance-Based.
Data Mining Classification: Alternative Techniques Lecture Notes for Chapter 5 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
Learning from observations
Data Mining Classification: Alternative Techniques Lecture Notes for Chapter 5 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar.
Data Mining Classification: Alternative Techniques Lecture Notes for Chapter 5 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar.
Outline K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
Lazy Learners K-Nearest Neighbor algorithm Fuzzy Set theory Classifier Accuracy Measures.
DATA MINING LECTURE 10b Classification k-nearest neighbor classifier
CS Machine Learning Instance Based Learning (Adapted from various sources)
K-Nearest Neighbor Learning.
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
Fall 2004, CIS, Temple University CIS527: Data Warehousing, Filtering, and Mining Lecture 8 Alternative Classification Algorithms Lecture slides taken/modified.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
Distance and Similarity Measures
Data Mining Classification: Alternative Techniques
Classification Nearest Neighbor
Instance Based Learning (Adapted from various sources)
K Nearest Neighbor Classification
Classification Nearest Neighbor
Nearest-Neighbor Classifiers
Instance Based Learning
COSC 4335: Other Classification Techniques
DATA MINING LECTURE 10 Classification k-nearest neighbor classifier
Nearest Neighbor Classifiers
Nearest Neighbors CSC 576: Data Mining.
Data Mining Classification: Alternative Techniques
Nearest Neighbor Classifiers
CSE4334/5334 Data Mining Lecture 7: Classification (4)
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Presentation transcript:

Navneet Goyal

Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)

Classification: Eager & Lazy Learners  Decision Tree classifier is an example of an “eager learner”  Because they are designed to learn a model that maps the input attributes to the class label as soon as the training data becomes available  An opposite strategy would be to delay the process of modeling the training data until it is needed to classify the test examples  LAZY Learners

Classification: Eager & Lazy Learners  Rote classifier is an example of lazy learner, which memorizes the entire training data & performs the classification only if the attributes of a test instance matches exactly with one of the training examples  Drawback: Cannot classify a new instance if it does match any training example

Classification: Nearest Neighbors  To overcome this drawback, we find all training examples that are relatively ‘similar’ to the test example  Examples, which are known as ‘Nearest Neighbors’ can be used to determine the class label of the test example  If it walks like a duck, quacks like a duck, and looks like a duck, then it is probably a duck

Nearest Neighbor Classifiers Training Records Test Record Compute Distance Choose k of the “nearest” records

Classification: Nearest Neighbors  Each test example is represented as a point in a d -dimensional space  For east test example we use a proximity measure  K-nearest neighbors of a given example z refer to the k points that are closest to z

Classification Using Distance  Place items in class to which they are “closest”.  Must determine distance between an item and a class.  Classes represented by  Centroid: Central value.  Medoid: Representative point.  Individual points  Algorithm: KNN

Definition of Nearest Neighbor K-nearest neighbors of a record x are data points that have the k smallest distance to x

Nearest-Neighbor Classifiers l Requires three things –The set of stored records –Distance Metric to compute distance between records –The value of k, the number of nearest neighbors to retrieve l To classify an unknown record: –Compute distance to other training records –Identify k nearest neighbors –Use class labels of nearest neighbors to determine the class label of unknown record (e.g., by taking majority vote)

Nearest Neighbor Classification  Compute distance between two points:  Euclidean distance Determine the class from nearest neighbor list  take the majority vote of class labels among the k-nearest neighbors  Weigh the vote according to distance  weight factor, w = 1/d 2

Nearest Neighbor Classification…  Choosing the value of k:  If k is too small, sensitive to noise points  If k is too large, neighborhood may include points from other classes

Nearest Neighbor Classification…  Problem with Euclidean measure:  High dimensional data  curse of dimensionality  Can produce counter-intuitive results vs d =  Solution: Normalize the vectors to unit length

Nearest neighbor Classification…  Part of more general technique called as ‘Instance-based learning’  Require a proximity measure  k-NN classifiers are lazy learners  It does not build models explicitly  Unlike eager learners such as decision tree induction and rule-based systems  Classifying unknown records are relatively expensive

Nearest neighbor Classification…  Make their classification based on local information, where as DT & rule-based classifiers attempt to find global model that fits the entire input space  As decisions are made locally, they are quite susceptible to noise (for small k)  Can produce wrong results unless the appropriate proximity measure and data preprocessing steps are taken

Nearest Neighbor Classification…  Scaling issues  Attributes may have to be scaled to prevent distance measures from being dominated by one of the attributes  Example:  height of a person may vary from 1.5m to 1.8m  weight of a person may vary from 90lb to 300lb  income of a person may vary from $10K to $1M  Proximity measure may be dominated by differences in weights and income of a person

Example: PEBLS  PEBLS: Parallel Examplar-Based Learning System (Cost & Salzberg)  Works with both continuous and nominal features  For nominal features, distance between two nominal values is computed using modified value difference metric (MVDM)  Each record is assigned a weight factor  Number of nearest neighbor, k = 1

Example: PEBLS Class Marital Status SingleMarriedDivorced Yes201 No241 Distance between nominal attribute values: d(Single,Married) = | 2/4 – 0/4 | + | 2/4 – 4/4 | = 1 d(Single,Divorced) = | 2/4 – 1/2 | + | 2/4 – 1/2 | = 0 d(Married,Divorced) = | 0/4 – 1/2 | + | 4/4 – 1/2 | = 1 d(Refund=Yes,Refund=No) = | 0/3 – 3/7 | + | 3/3 – 4/7 | = 6/7 Class Refund YesNo Yes03 No34

Example: PEBLS Distance between record X and record Y: where: w X  1 if X makes accurate prediction most of the time w X > 1 if X is not reliable for making predictions