Slides for “Data Mining” by I. H. Witten and E. Frank.

Slides:



Advertisements
Similar presentations
Rule-Based Classifiers. Rule-Based Classifier Classify records by using a collection of “if…then…” rules Rule: (Condition)  y –where Condition is a conjunctions.
Advertisements

CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Decision Trees with Numeric Tests
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Classification Algorithms – Continued. 2 Outline  Rules  Linear Models (Regression)  Instance-based (Nearest-neighbor)
Classification Algorithms – Continued. 2 Outline  Rules  Linear Models (Regression)  Instance-based (Nearest-neighbor)
Decision Tree Learning 主講人:虞台文 大同大學資工所 智慧型多媒體研究室.
Decision Trees IDHairHeightWeightLotionResult SarahBlondeAverageLightNoSunburn DanaBlondeTallAverageYesnone AlexBrownTallAverageYesNone AnnieBlondeShortAverageNoSunburn.
Algorithms: The basic methods. Inferring rudimentary rules Simplicity first Simple algorithms often work surprisingly well Many different kinds of simple.
Knowledge Representation. 2 Outline: Output - Knowledge representation  Decision tables  Decision trees  Decision rules  Rules involving relations.
Decision Tree Algorithm
Induction of Decision Trees
Tree-based methods, neutral networks
Knowledge Representation. 2 Outline: Output - Knowledge representation  Decision tables  Decision trees  Decision rules  Rules involving relations.
Covering Algorithms. Trees vs. rules From trees to rules. Easy: converting a tree into a set of rules –One rule for each leaf: –Antecedent contains a.
Classification.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Data Mining – Output: Knowledge Representation
Machine Learning Chapter 3. Decision Tree Learning
Inductive learning Simplest form: learn a function from examples
Mohammad Ali Keyvanrad
Figure 1.1 Rules for the contact lens data.. Figure 1.2 Decision tree for the contact lens data.
Learning from Observations Chapter 18 Through
Data Mining Practical Machine Learning Tools and Techniques Chapter 3: Output: Knowledge Representation Rodney Nielsen Many of these slides were adapted.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.3: Decision Trees Rodney Nielsen Many of.
Data Management and Database Technologies 1 DATA MINING Extracting Knowledge From Data Petr Olmer CERN
Decision Trees. MS Algorithms Decision Trees The basic idea –creating a series of splits, also called nodes, in the tree. The algorithm adds a node to.
Multi-Relational Data Mining: An Introduction Joe Paulowskey.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.7: Instance-Based Learning Rodney Nielsen.
CS690L Data Mining: Classification
1 CSC 9010 Spring Paula Matuszek CSC 9010 ANN Lab Paula Matuszek Spring, 2011.
Decision Trees. What is a decision tree? Input = assignment of values for given attributes –Discrete (often Boolean) or continuous Output = predicated.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.2 Statistical Modeling Rodney Nielsen Many.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Sections 4.1 Inferring Rudimentary Rules Rodney Nielsen.
Daniel Kroening and Ofer Strichman 1 Decision Procedures An Algorithmic Point of View BDDs.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall 6.8: Clustering Rodney Nielsen Many / most of these.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.2: Classification Rules Rodney Nielsen Many.
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 3 of Data Mining by I. H. Witten, E. Frank and M. A. Hall.
CIS 335 CIS 335 Data Mining Classification Part I.
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Linear Models & Clustering Presented by Kwak, Nam-ju 1.
Data Mining Chapter 4 Algorithms: The Basic Methods Reporter: Yuen-Kuei Hsueh.
CSE573 Autumn /09/98 Machine Learning Administrative –Last topic: Decision Tree Learning Reading: 5.1, 5.4 Last time –finished NLP sample system’s.
Rodney Nielsen Many of these slides were adapted from: I. H. Witten, E. Frank and M. A. Hall Data Science Output: Knowledge Representation WFH: Data Mining,
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 3 of Data Mining by I. H. Witten and E. Frank.
Data Mining Practical Machine Learning Tools and Techniques
Data Science Algorithms: The Basic Methods
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Data Science Algorithms: The Basic Methods
Rule-Based Classification
Data Science Algorithms: The Basic Methods
CSE 711: DATA MINING Sargur N. Srihari Phone: , ext. 113.
Decision Tree Saed Sayad 9/21/2018.
Figure 1.1 Rules for the contact lens data.
Data Mining for Business Analytics
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning Chapter 3. Decision Tree Learning
Machine Learning in Practice Lecture 17
Data Mining CSCI 307, Spring 2019 Lecture 6
Data Mining CSCI 307, Spring 2019 Lecture 9
Presentation transcript:

Slides for “Data Mining” by I. H. Witten and E. Frank

2 Output: Knowledge representation  Decision tables  Decision trees  Decision rules  Association rules  Rules with exceptions  Rules involving relations  Linear regression  Trees for numeric prediction  Instance-based representation  Clusters 3

3 Output: representing structural patterns  Many different ways of representing patterns  Decision trees, rules, instance-based, …  Also called “knowledge” representation  Representation determines inference method  Understanding the output is the key to understanding the underlying learning methods  Different types of output for different learning problems (e.g. classification, regression, …)

4 Decision tables  Simplest way of representing output:  Use the same format as input!  Decision table for the weather problem:  Main problem: selecting the right attributes OutlookHumidityPlay SunnyHighNo SunnyNormalYes OvercastHighYes OvercastNormalYes RainyHighNo RainyNormalNo

5 Decision trees  “Divide-and-conquer” approach produces tree  Nodes involve testing a particular attribute  Usually, attribute value is compared to constant  Other possibilities:  Comparing values of two attributes  Using a function of one or more attributes  Leaves assign classification, set of classifications, or probability distribution to instances  Unknown instance is routed down the tree

6 Nominal and numeric attributes  Nominal: number of children usually equal to number values  attribute won’t get tested more than once  Other possibility: division into two subsets  Numeric: test whether value is greater or less than constant  attribute may get tested several times  Other possibility: three-way split (or multi-way split)  Integer: less than, equal to, greater than  Real: below, within, above

7 Missing values  Does absence of value have some significance?  Yes  “missing” is a separate value  No  “missing” must be treated in a special way  Solution A: assign instance to most popular branch  Solution B: split instance into pieces  Pieces receive weight according to fraction of training instances that go down each branch  Classifications from leave nodes are combined using the weights that have percolated to them

8 Classification rules  Popular alternative to decision trees  Antecedent (pre-condition): a series of tests (just like the tests at the nodes of a decision tree)  Tests are usually logically ANDed together (but may also be general logical expressions)  Consequent (conclusion): classes, set of classes, or probability distribution assigned by rule  Individual rules are often logically ORed together  Conflicts arise if different conclusions apply

9 From trees to rules  Easy: converting a tree into a set of rules  One rule for each leaf:  Antecedent contains a condition for every node on the path from the root to the leaf  Consequent is class assigned by the leaf  Produces rules that are unambiguous  Doesn’t matter in which order they are executed  But: resulting rules are unnecessarily complex  Pruning to remove redundant tests/rules

10 From rules to trees  More difficult: transforming a rule set into a tree  Tree cannot easily express disjunction between rules  Example: rules which test different attributes  Symmetry needs to be broken  Corresponding tree contains identical subtrees (  “replicated subtree problem”) If a and b then x If c and d then x

11 A tree for a simple disjunction

12 The exclusive-or problem If x = 1 and y = 0 then class = a If x = 0 and y = 1 then class = a If x = 0 and y = 0 then class = b If x = 1 and y = 1 then class = b

13 A tree with a replicated subtree If x = 1 and y = 1 then class = a If z = 1 and w = 1 then class = a Otherwise class = b

14 “Nuggets” of knowledge  Are rules independent pieces of knowledge? (It seems easy to add a rule to an existing rule base.)  Problem: ignores how rules are executed  Two ways of executing a rule set:  Ordered set of rules (“decision list”)  Order is important for interpretation  Unordered set of rules  Rules may overlap and lead to different conclusions for the same instance

15 Interpreting rules  What if two or more rules conflict?  Give no conclusion at all?  Go with rule that is most popular on training data?  …  What if no rule applies to a test instance?  Give no conclusion at all?  Go with class that is most frequent in training data?  …

16 Special case: boolean class  Assumption: if instance does not belong to class “yes”, it belongs to class “no”  Trick: only learn rules for class “yes” and use default rule for “no”  Order of rules is not important. No conflicts!  Rule can be written in disjunctive normal form If x = 1 and y = 1 then class = a If z = 1 and w = 1 then class = a Otherwise class = b

17 Rules involving relations  So far: all rules involved comparing an attribute-value to a constant (e.g. temperature < 45)  These rules are called “propositional” because they have the same expressive power as propositional logic  What if problem involves relationships between examples (e.g. family tree problem from above)?  Can’t be expressed with propositional rules  More expressive representation required

18 The shapes problem  Target concept: standing up  Shaded: standing Unshaded: lying

19 A propositional solution WidthHeightSidesClass 244Standing Lying 783Standing 763Lying 294Standing 914Lying 1023Lying If width  3.5 and height < 7.0 then lying If height  3.5 then standing

20 A relational solution  Comparing attributes with each other  Generalizes better to new data  Standard relations: =,  But: learning relational rules is costly  Simple solution: add extra attributes (e.g. a binary attribute is width < height?) If width > height then lying If height > width then standing

21 Rules with variables  Using variables and multiple relations:  The top of a tower of blocks is standing:  The whole tower is standing:  Recursive definition! If height_and_width_of(x,h,w) and h > w then standing(x) If is_top_of(x,z) and height_and_width_of(z,h,w) and h > w and is_rest_of(x,y)and standing(y) then standing(x) If empty(x) then standing(x) If height_and_width_of(x,h,w) and h > w and is_top_of(x,y) then standing(x)

22 Inductive logic programming  Recursive definition can be seen as logic program  Techniques for learning logic programs stem from the area of “inductive logic programming” (ILP)  But: recursive definitions are hard to learn  Also: few practical problems require recursion  Thus: many ILP techniques are restricted to non-recursive definitions to make learning easier