Budowa reguł decyzyjnych z rozmytą granulacją wiedzy Zenon A. Sosnowski Wydział Informatyki Politechnika Białostocka Wiejska 45A, 15-351 Bialystok

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Clustering Clustering of data is a method by which large sets of data is grouped into clusters of smaller sets of similar data. The example below demonstrates.
DECISION TREES. Decision trees  One possible representation for hypotheses.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Hunt’s Algorithm CIT365: Data Mining & Data Warehousing Bajuna Salehe
Decision Tree Approach in Data Mining
Classification: Definition Given a collection of records (training set ) –Each record contains a set of attributes, one of the attributes is the class.
1 Data Mining Classification Techniques: Decision Trees (BUSINESS INTELLIGENCE) Slides prepared by Elizabeth Anglo, DISCS ADMU.
Decision Tree.
1 Machine Learning: Lecture 10 Unsupervised Learning (Based on Chapter 9 of Nilsson, N., Introduction to Machine Learning, 1996)
K Means Clustering , Nearest Cluster and Gaussian Mixture
Clustering… in General In vector space, clusters are vectors found within  of a cluster vector, with different techniques for determining the cluster.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
© University of Minnesota Data Mining for the Discovery of Ocean Climate Indices 1 CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance.
Induction of Decision Trees
Learning From Data Chichang Jou Tamkang University.
Lecture 5 (Classification with Decision Trees)
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
ROUGH SET THEORY AND FUZZY LOGIC BASED WAREHOUSING OF HETEROGENEOUS CLINICAL DATABASES Yiwen Fan.
Evaluating Performance for Data Mining Techniques
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
A Simple Method to Extract Fuzzy Rules by Measure of Fuzziness Jieh-Ren Chang Nai-Jian Wang.
1 Data Mining Lecture 3: Decision Trees. 2 Classification: Definition l Given a collection of records (training set ) –Each record contains a set of attributes,
Chapter 10. Sampling Strategy for Building Decision Trees from Very Large Databases Comprising Many Continuous Attributes Jean-Hugues Chauchat and Ricco.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
Universit at Dortmund, LS VIII
1 Motivation Web query is usually two or three words long. –Prone to ambiguity –Example “keyboard” –Input device of computer –Musical instruments How can.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
1 Learning Chapter 18 and Parts of Chapter 20 AI systems are complex and may have many parameters. It is impractical and often impossible to encode all.
1 Statistical Techniques Chapter Linear Regression Analysis Simple Linear Regression.
Witold Pedrycz Department of Electrical & Computer Engineering University of Alberta, Edmonton, Canada and Systems Research Institute, Polish Academy of.
ID3 Algorithm Michael Crawford.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Data Mining and Decision Trees 1.Data Mining and Biological Information 2.Data Mining and Machine Learning Techniques 3.Decision trees and C5 4.Applications.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Linear Models for Classification
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
Algorithmic Facets of Human Centricity in Computing with Fuzzy Sets ISDA-2009, Pisa, Italy, November 30-December 2, 2009
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Apache Mahout Qiaodi Zhuang Xijing Zhang.
Lecture Notes for Chapter 4 Introduction to Data Mining
Recent Trends in Fuzzy Clustering: From Data to Knowledge Shenyang, August 2009
1 Classification: predicts categorical class labels (discrete or nominal) classifies data (constructs a model) based on the training set and the values.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
DECISION TREES Asher Moody, CS 157B. Overview  Definition  Motivation  Algorithms  ID3  Example  Entropy  Information Gain  Applications  Conclusion.
Identifying Ethnic Origins with A Prototype Classification Method Fu Chang Institute of Information Science Academia Sinica ext. 1819
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Fuzzy Pattern Recognition. Overview of Pattern Recognition Pattern Recognition Procedure Feature Extraction Feature Reduction Classification (supervised)
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Non-parametric Methods for Clustering Continuous and Categorical Data Steven X. Wang Dept. of Math. and Stat. York University May 13, 2010.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
CLUSTER ANALYSIS. Cluster Analysis  Cluster analysis is a major technique for classifying a ‘mountain’ of information into manageable meaningful piles.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Algorithms: The Basic Methods Clustering WFH:
Fuzzy Logic in Pattern Recognition
Clustering CSC 600: Data Mining Class 21.
Privacy-Preserving Data Mining
Decision Trees.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Fast Kernel-Density-Based Classification and Clustering Using P-Trees
Basic machine learning background with Python scikit-learn
Clustering (3) Center-based algorithms Fuzzy k-means
Fall Risk Assessment.
Unsupervised Optimal Fuzzy Clustering
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Clustering Wei Wang.
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Boltzmann Machine (BM) (§6.4)
Presentation transcript:

Budowa reguł decyzyjnych z rozmytą granulacją wiedzy Zenon A. Sosnowski Wydział Informatyki Politechnika Białostocka Wiejska 45A, Bialystok

Agenda wprowadzenie drzewa decyzyjne (DT) zbiory rozmyte w granulacji atrybutów algorytm generowania kontekstowych DT przykład wnioski

Rozmyta sieć RETE The inference mechanism realizes a generalized modus ponens rule. if A then C CFr A' CFf C'CFc CFr is an uncertainty of the rule CFf is an uncertainty of the fact CFc is an uncertainty of the conclusion CFc = CFr * CFf

Fuzzy_Fuzzy

(defrule r1 (speed very fast) => (... )) (defrule r2 (speed slow) => (... )) SINGLE (LV speed) MULTIFIELD End of pattern activation rule r2 M.(slow) (attached) M.(very fast) (attached) (speed medium) - WME

Decicion Trees – An Overview used to solve classification problems structure of problem - attributes - each attribute assumes a finite number values - finite number of discrete classes entropy-based optimization criterion architecture of decision tree: nodes – attributes, edges – values of attributes

Coping with Continuous Attributes Decision trees require finite-valued attributes What if attributes are continuous ? Attributes need to be discretrized Options: - discretize each attribute separately (uniform and nonuniform) - discretize all attributes (clustering )

Quantization of attributes through clustering Fuzzy Clustering Context-based fuzzy clustering

Fuzzy Clustering (FCM) versus Context-Based FCM (cFCM) Fuzzy clustering: objective function and its iteraive optimization Context-base fuzzy clustering: - objective function minimized iteratively - continuous classification variable granulated with the use of linguistic labels

Context-Based Fuzzy Clustering Given: data {x k,y k }, k=1,2,…,N, number of clusters (c), distance function ||.||, fuzzy set of context A defined over y k Constrained-based optimization of objective function subject to

From context fuzzy set A to the labeling of data to be clustered

Context-Based Fuzzy Clustering: An Iterative Optimization Process Given: The number of clusters (c). Select the distance function ||.||, termination criterion e (>0) and initialize partition matrix U U. Select the value of the fuzzification parameter “m” (the default is m=2.0) 1.Calculate centers (prototypes) of the clusters i=1, 2,..., c 2. Update partition matrix i=1, 2,..., c, j=1, 2,..., N 3. Compare U' to U, if termination criterion ||U’ - U|| <e is satisfied then stop, else return to step (1) and proceed with computing by setting up U equal to U' Result: partition matrix and prototypes

Information Granules in the Development of Decision Trees define contexts (fuzzy sets) for continuous classivication variable cluster data for each context project prototypes on the individual axes – this leads to their discretization carry out the standard ID-3 algorithm W. Pedrycz, Z.A. Sosnowski, „The designing of decision trees in the framework of granular data and their application to software quality models”, Fuzzy Sets & Sysytems, vol. 124, (2001), p

Fuzzy Sets of Contexts: Two Approaches subjective selection depending on the classification problem supported by statistical relevance (σ-count of fuzzy contexts)

Constructing linguistic terms – classes (thin line) and their induced interval- valued counterparts (solid line)

C - Fuzzy Decision Trees W. Pedrycz, Z.A. Sosnowski, „C-Fuzzy Decision Trees”, IEEE Transactions on Systems, Man and Cybernetics, Part C, Vol. 35, No 4, 2005, p

Architecture of the cluster-based decision tree cluster all data set X repeat allocate elements of X to each cluster choose the node with the highest value of the spliting criterion cluster data at selected node until termination criterion is fulfield

Node splitting criterion Node of the tree Ni = where: Xi = { x(k) | ui(x(k)) > uj(x(k))} Yi = {y(k)| x(k) ε Xi} Ui = [ui(x(1)) ui(x(2)) … ui(x(N))]

Stopping criterion (structurability index)

C-fuzzy tree in the classification (prediction) mode assign x to class w i if u i (x) exceeds the values of the membership in all remaining clusters

Experiments Data sets from the UCI repository of Machine Learning Databases ( Auto-Mpg Pima-diabetes Ionosphere Hepatitis Dermatology

Hepatitis data Type of tree and its structural parameters Error: Training data Error: Testing dataNumber of nodes C4.5 rev % (average) 0.85 % (st. deviation) % (average) 7.05 % (st. deviation) 45 (average) 7.87 (st. deviation) C-decision tree c=2 clusters, 6 iterations % (average) 3.34 % (st. deviation) % (average) 0.08 % (st. deviation) 12 C-decision tree c=9 clusters, 3 iterations % (average) 5.21 % (st. deviation) % (average) 3.68 % (st. deviation) 27

Dermatology data Type of tree and its structural parameters Error: Training data Error: Testing dataNumber of nodes C4.5 rev % (average) 0.61 % (st. deviation) 5,98% (average) 3.50% (st. deviation) 18.6 (average) 4.34 (st. deviation) C-decision tree C=11 clusters, 1 iterations 7.0 % (average) 1,68 % (st. deviation) 4.9 % (average) 3.56 % (st. deviation) 11 C-decision tree c=7 clusters, 1 iterations 6.1 % (average) 1.15 % (st. deviation) 5.7 % (average) 2.47 % (st. deviation) 7

Context-based Fuzzy Clustered- oriented Decision Trees (CFCDT).....

Architecture of the Context-based Fuzzy Clustered-oriented Decision Tree define contexts (fuzzy sets) for classivication variable for each context do –cluster (cFCM) X i (data set of i-th context) –repeat –allocate elements of X i to each cluster –choose the node with the highest value of the spliting criterion –cluster (cFCM) data at selected node until termination criterion is fulfield enddo

Problem Implementation issues: high complexity –> grid or cluster computing agregation -> testing of different appraches

Dziękuję za uwagę