Download presentation
Presentation is loading. Please wait.
Published byEverett McKinney Modified over 9 years ago
1
Data Mining By Farzana Forhad CS 157B
2
Agenda Decision Tree and ID3 Rough Set Theory Clustering
3
Introduction Data mining is a component of a wider process called knowledge discovery from databases. Data mining is a component of a wider process called knowledge discovery from databases. The basic foundations of data mining: The basic foundations of data mining: –decision tree –association rules –clustering –other statistical techniques
4
Decision Tree ID3 (Quinlan 1986), represents concepts as decision trees. ID3 (Quinlan 1986), represents concepts as decision trees. A decision tree is a classifier in the form of a tree structure where each node is either: A decision tree is a classifier in the form of a tree structure where each node is either: –a leaf node, indicating a class of instances OR –a decision node, which specifies a test to be carried out on a single attribute value, with one branch and a sub-tree for each possible outcome of the test
5
Decision Tree The set of records available for classification is divided into two disjoint subsets: The set of records available for classification is divided into two disjoint subsets: –a training set : used for deriving the classifier –a test set: used to measure the accuracy of the classifier Attributes whose domain is numerical are called numerical attributes Attributes whose domain is not numerical are called categorical attributes.
6
Decision Tree A decision tree is a tree with the following properties: A decision tree is a tree with the following properties: –An inner node represents an attribute –An edge represents a test on the attribute of the father node –A leaf represents one of the classes Construction of a decision tree Construction of a decision tree –Based on the training data –Top-Down strategy
7
Training Dataset
8
Test Dataset
9
Decision Tree RULE 1 If it is sunny and the humidity is not above 75%, then play. RULE 2 If it is sunny and the humidity is above 75%, then do not play. RULE 3 If it is overcast, then play. RULE 4 If it is rainy and not windy, then play. RULE 5 If it is rainy and windy, then don't play.
10
Training Dataset
11
Decision Tree for Zip Code and Age
12
Iterative Dichotomizer 3 (ID3) Quinlan (1986) Quinlan (1986) Each node corresponds to a splitting attribute Each node corresponds to a splitting attribute –Entropy is used to measure how informative is a node. –The algorithm uses the criterion of information gain to determine the goodness of a split.
13
Iterative Dichotomizer 3 (ID3)
14
Rough Set Theory –Useful means for studying delivery patterns, rules, and knowledge in data –The rough set is the estimate of a vague concept by a pair of specific concepts, called the lower and upper approximations.
15
Rough Set Theory –The lower approximation is a type of the domain objects which are known with certainty to belong to the subset of interest. – The upper approximation is a description of the objects which may perhaps belong to the subset. –Any subset defined through its lower and upper approximations is called a rough set, if the boundary region is not empty.
16
Lower and Upper Approximations of a Rough Set
17
Association Rule Mining Basket Analysis Basket Analysis
18
Definition of Association Rules
19
Mining the Rules
20
Two Steps of Association Rule Mining
21
Clustering Clustering The process of organizing objects into groups whose members are similar in some way The process of organizing objects into groups whose members are similar in some way Statistics, machine learning, and database researchers have studied data clustering Statistics, machine learning, and database researchers have studied data clustering Recent emphasis on large datasets Recent emphasis on large datasets
22
Different Approaches to Clustering Two main approaches to clustering: Two main approaches to clustering: -partitioning clustering -hierarchical clustering Clustering algorithms differ among themselves in the following ways: Clustering algorithms differ among themselves in the following ways: –in their ability to handle different types of attributes (numeric and categorical) –in accuracy of clustering –in their ability to handle disk-resident data
23
Problem Statement N objects to be grouped in k clusters N objects to be grouped in k clusters Number of different possibilities: Number of different possibilities: The objective is to find a grouping such that the distances between objects in a group is minimum The objective is to find a grouping such that the distances between objects in a group is minimum Several algorithms to find near optimal solution Several algorithms to find near optimal solution
24
k-Means Algorithm 1. Randomly select k points to be the starting points for the centroids of the k clusters. 2. Assign each object to the centroid closest to the object, forming k exclusive clusters of examples. 3. Calculate new centroids of the clusters. Take the average of all the attribute values of the objects belonging to the same cluster. 4. Check if the cluster centroids have changed their coordinates. If yes, repeat from Step 2. 5. If no, cluster detection is finished, and all objects have their cluster memberships defined.
25
Example One-dimensional database with N = 9 One-dimensional database with N = 9 Objects labeled z 1 …z 9 Objects labeled z 1 …z 9 Let k = 2 Let k = 2 Let us start with z 1 to z 2 as the initial centroids Let us start with z 1 to z 2 as the initial centroids Table: One- dimensional database
26
Example Table: New cluster assignments
27
Example Table: Reassignment of objects to two clusters
28
Questions? Thank You
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.