Identifying Competence-Critical Instances for Instance-Based Learners 2001. 5. 9 Presenter: Kyu-Baek Hwang.

Slides:



Advertisements
Similar presentations
Case Based Reasoning Lecture 7: CBR Competence of Case-Bases.
Advertisements

INSTANCE-BASED LEARNING ALGORITHMS Presented by Yan T. Yang.
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
Yue Han and Lei Yu Binghamton University.
Instance Based Learning
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
Lazy vs. Eager Learning Lazy vs. eager learning
5. 1 Model of Image degradation and restoration
Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised)
K nearest neighbor and Rocchio algorithm
Mutual Information Mathematical Biology Seminar
Carla P. Gomes CS4700 CS 4700: Foundations of Artificial Intelligence Carla P. Gomes Module: Nearest Neighbor Models (Reading: Chapter.
Decision Tree Algorithm
Feature Selection for Regression Problems
Instance Based Learning. Nearest Neighbor Remember all your data When someone asks a question –Find the nearest old data point –Return the answer associated.
Instance Based Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán) 1.
Nearest Neighbour Condensing and Editing David Claus February 27, 2004 Computer Vision Reading Group Oxford.
INSTANCE-BASE LEARNING
Applications of Data Mining in Microarray Data Analysis Yen-Jen Oyang Dept. of Computer Science and Information Engineering.
CS Instance Based Learning1 Instance Based Learning.
Simple search methods for finding a Nash equilibrium Ryan Porter, Eugene Nudelman, and Yoav Shoham Games and Economic Behavior, Vol. 63, Issue 2. pp ,
Module 04: Algorithms Topic 07: Instance-Based Learning
ENN: Extended Nearest Neighbor Method for Pattern Recognition
1 Lazy Learning – Nearest Neighbor Lantz Ch 3 Wk 2, Part 1.
Processing of large document collections Part 2 (Text categorization) Helena Ahonen-Myka Spring 2006.
Dealing with Noisy Data. 1.Introduction 2.Identifying Noise 3.Types of Noise 4.Noise Filtering 5.Robust Learners Against Noise 6.Experimental Comparative.
Case Base Maintenance(CBM) Fabiana Prabhakar CSE 435 November 6, 2006.
Instance Selection. 1.Introduction 2.Training Set Selection vs. Prototype Selection 3.Prototype Selection Taxonomy 4.Description of Methods 5.Related.
Theory Revision Chris Murphy. The Problem Sometimes we: – Have theories for existing data that do not match new data – Do not want to repeat learning.
CS Learning Rules1 Learning Sets of Rules. CS Learning Rules2 Learning Rules If (Color = Red) and (Shape = round) then Class is A If (Color.
Basic Data Mining Technique
K Nearest Neighbors Saed Sayad 1www.ismartsoft.com.
CSC 196k Semester Project: Instance Based Learning
MA4102 – Data Mining and Neural Networks Nathan Ifill University of Leicester Image source: Antti Ajanki, “Example of k-nearest neighbor.
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
1 Instance Based Learning Ata Kaban The University of Birmingham.
David Claus and Christoph F. Eick: Nearest Neighbor Editing and Condensing Techniques Nearest Neighbor Editing and Condensing Techniques 1.Nearest Neighbor.
1 Redundant Feature Elimination for Multi-Class Problems Annalisa Appice, Michelangelo Ceci Dipartimento di Informatica, Università degli Studi di Bari,
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.7: Instance-Based Learning Rodney Nielsen.
1 Eick, Zeidat, Vilalta: Using Representative-based Clustering for NN Dataset Editing (ICDM04) Using Representative-Based Clustering For Nearest Neighbour.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. A fast nearest neighbor classifier based on self-organizing incremental neural network (SOINN) Neuron.
On the Role of Dataset Complexity in Case-Based Reasoning Derek Bridge UCC Ireland (based on work done with Lisa Cummins)
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
Data Mining: Knowledge Discovery in Databases Peter van der Putten ALP Group, LIACS Pre-University College LAPP-Top Computer Science February 2005.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
1 Estimating the reliability of the kNN classification Maxim Tsypin and Heinrich Röder Biodesix, Steamboat Springs, CO 1.kNN classification 2.Problems.
An Approximate Nearest Neighbor Retrieval Scheme for Computationally Intensive Distance Measures Pratyush Bhatt MS by Research(CVIT)
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.2: Classification Rules Rodney Nielsen Many.
COMP 2208 Dr. Long Tran-Thanh University of Southampton K-Nearest Neighbour.
On Utillizing LVQ3-Type Algorithms to Enhance Prototype Reduction Schemes Sang-Woon Kim and B. John Oommen* Myongji University, Carleton University*
1 Software Testing and Quality Assurance Lecture 17 - Test Analysis & Design Models (Chapter 4, A Practical Guide to Testing Object-Oriented Software)
CS Machine Learning Instance Based Learning (Adapted from various sources)
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Advisor : Dr. Hsu Presenter : Chien-Shing Chen Author: Gustavo.
Enhanced hypertext categorization using hyperlinks Soumen Chakrabarti (IBM Almaden) Byron Dom (IBM Almaden) Piotr Indyk (Stanford)
Eick: kNN kNN: A Non-parametric Classification and Prediction Technique Goals of this set of transparencies: 1.Introduce kNN---a popular non-parameric.
SEMI-SYNTHETIC CIRCUIT GENERATION FOR TESTING INCREMENTAL PLACE AND ROUTE TOOLS David GrantGuy Lemieux University of British Columbia Vancouver, BC.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Instance-Based Learning Evgueni Smirnov. Overview Instance-Based Learning Comparison of Eager and Instance-Based Learning Instance Distances for Instance-Based.
Data Mining Practical Machine Learning Tools and Techniques Chapter 6.5: Instance-based Learning Rodney Nielsen Many / most of these slides were adapted.
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss Pedro Domingos, Michael Pazzani Presented by Lu Ren Oct. 1, 2007.
CS 326A: Motion Planning Probabilistic Roadmaps for Path Planning in High-Dimensional Configuration Spaces (1996) L. Kavraki, P. Švestka, J.-C. Latombe,
Information Management course
K Nearest Neighbors and Instance-based methods
K Nearest Neighbor Classification
Goodfellow: Chapter 14 Autoencoders
CSE572, CBS572: Data Mining by H. Liu
CSE572: Data Mining by H. Liu
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

Identifying Competence-Critical Instances for Instance-Based Learners Presenter: Kyu-Baek Hwang

Abstract The basic nearest neighbor classifier with a large dataset Classification accuracy and response time Review on past works tackling these problems No consistent method Insight into the problem characteristics Iterative case filtering (ICF) algorithm

Introduction Harmful and superfluous instances are stored. Selectively store instances (or delete stored instances) The data miner have to gain an insight into the structure of the classes in the instance space. The experimental comparison of RT3 and ICF Neither algorithm performs better in all cases.

Defining the Problem Two practical issues that arise in this area  Instance removal (retain only the critical instances)  Different approaches according to the type of the classification problem The same (or higher) accuracy and the less storage Which instance should be deleted?

Four Cases Where NNC Fails Noisy instance Close to the interclass border  Border instances are critical in general. Small region defining the class  Small k values cope with this kind of problem. Unsolvable problem

Instance Space Structure Two categories of instance space structure  Homogeneous region (locality)  Non-homogeneous region (no locality)

Which Instances Are Critical? Prototypes  For non-homogeneous regions Instances with high utility  Needs classification feedback Instances which lie on borders are almost always critical.

Review Competence enhancement  By removing noisy or corrupt instances Competence preservation  By removing superfluous instances Hybrid approach  Many modern approaches

Competence Enhancement Stochastic noise Wilson Editing  All instances which are incorrectly classified by their nearest neighbors are assumed to be nosy instances.  Smoothing effect  Empirically tested Noisy instances and genuine exceptions

Competence Preservation Condensed nearest neighbor (CNN)  Look for cases for which removal does not lead to additional miss- classification Chang’s algorithm (Korean)  Merging two instances into one synthetic instance (the prototype) Footprint deletion policy  Local-set of a case c  The set of cases contained in the largest hypersphere centered on c such that only cases in the same class as c are contained in the hypersphere.

Footprint Deletion Policy For a case-base CB = {c 1, c 2, …, c n }  Coverage(c) = {c’  CB: c’  Local-set(c)}  Reachable(c) = {c’  CB: c  Local-set(c’)} Pivotal group  With an empty reachable set Delete the instance with large local-set

Hybrid Approaches (1/2) IB2 (on-line)  If a new case to be added can already be classified correctly on the basis of the current case-base, the case is discarded. IB3  IB2 with time delay The order of presentation is crucial for IB2 and IB3. RT1  k nearest neighbor  Associates of the case p are the cases that have p as their k nearest neighbor.  The instance which has many associates is tested and removed.

Hybrid Approaches (2/2) RT2 is identical to RT1 and additionally,  Cases furthest from their nearest enemy are removed first.  Removed associates still guide the deletion process. RT3 is identical to RT2 and additionally,  Wilson’s noise filtering step is executed first. RT algorithms are analogous to the footprint deletion policy.

An Iterative Case Filtering Algorithm Coverage set and reachable set RTn algorithm  Associate set of fixed size Remove cases which have a reachable set size greater than the coverage set size.  Intuitively, this approach removes the cases that are far from the border. A noisy case will have a singleton reachable set and a singleton coverage set.  This property protects the noisy case from being removed.  Wilson Editing

ICF Algorithm

How The ICF Algorithm Proceeds?

Experiments Experiments on 30 datasets from UCI repository Maximum number of iterations: 17  switzerland database In general, 3 iterations are required.

Reduction Profiles The percentage of cases removed after each iteration  switzerland database: 17 iterations, 2 – 13% (complicated)  zoo database: 2 iterations, 37% (simple structure)

Comparative Evaluation (1) Early approaches  CNN, RNN, SNN, Chang, Wilson Editing, repeated Wilson Editing, and all k-NN (2) Recent editions  IB2, IB3, TIBLE, and IBL-MDL (3) State of the art  RT3 and ICF

RT3 and ICF

Conclusions The structure of the instance space is important. ICF and RT3 behave in very similar way.  The intrinsic properties of them are similar.  80% of removal and the little degradation of accuracy. The reduction profile provides some insights into the property of the problem.