CS4445/B12 Provided by: Kenneth J. Loomis. CLASSIFICATION RULES: RIPPER ALGORITHM.

Slides:



Advertisements
Similar presentations
Data Mining Classification: Alternative Techniques
Advertisements

Mining Association Rules
Recap: Mining association rules from large datasets
Huffman Codes and Asssociation Rules (II) Prof. Sin-Min Lee Department of Computer Science.
Association Analysis (2). Example TIDList of item ID’s T1I1, I2, I5 T2I2, I4 T3I2, I3 T4I1, I2, I4 T5I1, I3 T6I2, I3 T7I1, I3 T8I1, I2, I3, I5 T9I1, I2,
Frequent Itemset Mining Methods. The Apriori algorithm Finding frequent itemsets using candidate generation Seminal algorithm proposed by R. Agrawal and.
DATA MINING Association Rule Discovery. AR Definition aka Affinity Grouping Common example: Discovery of which items are frequently sold together at a.
Data Mining (Apriori Algorithm)DCS 802, Spring DCS 802 Data Mining Apriori Algorithm Spring of 2002 Prof. Sung-Hyuk Cha School of Computer Science.
RIPPER Fast Effective Rule Induction
Association Rule Mining. 2 The Task Two ways of defining the task General –Input: A collection of instances –Output: rules to predict the values of any.
Data Mining Association Analysis: Basic Concepts and Algorithms
Association Mining Data Mining Spring Transactional Database Transaction – A row in the database i.e.: {Eggs, Cheese, Milk} Transactional Database.
Association Rules l Mining Association Rules between Sets of Items in Large Databases (R. Agrawal, T. Imielinski & A. Swami) l Fast Algorithms for.
Rakesh Agrawal Ramakrishnan Srikant
Association Analysis. Association Rule Mining: Definition Given a set of records each of which contain some number of items from a given collection; –Produce.
Data Mining Techniques So Far: Cluster analysis K-means Classification Decision Trees J48 (C4.5) Rule-based classification JRIP (RIPPER) Logistic Regression.
Data Mining Association Analysis: Basic Concepts and Algorithms Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach, Kumar Introduction.
Data Mining Association Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 6 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach,
Data Mining Association Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 6 Introduction to Data Mining by Tan, Steinbach, Kumar © Tan,Steinbach,
Mining Association Rules. Association rules Association rules… –… can predict any attribute and combinations of attributes … are not intended to be used.
Association Analysis (2). Example TIDList of item ID’s T1I1, I2, I5 T2I2, I4 T3I2, I3 T4I1, I2, I4 T5I1, I3 T6I2, I3 T7I1, I3 T8I1, I2, I3, I5 T9I1, I2,
Association Rule Mining Part 2 (under construction!) Introduction to Data Mining with Case Studies Author: G. K. Gupta Prentice Hall India, 2006.
Data Mining Association Analysis: Basic Concepts and Algorithms
Data Quality Class 9. Rule Discovery Decision and Classification Trees Association Rules.
ID3 Algorithm Abbas Rizvi CS157 B Spring What is the ID3 algorithm? ID3 stands for Iterative Dichotomiser 3 Algorithm used to generate a decision.
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
Association Analysis: Basic Concepts and Algorithms.
Association Rule Mining. Generating assoc. rules from frequent itemsets  Assume that we have discovered the frequent itemsets and their support  How.
6/23/2015CSE591: Data Mining by H. Liu1 Association Rules Transactional data Algorithm Applications.
Association Action Rules b y Zbigniew W. Ras 1,5 Agnieszka Dardzinska 2 Li-Shiang Tsay 3 Hanna Wasyluk 4 1)University of North Carolina, Charlotte, NC,
© Vipin Kumar CSci 8980 Fall CSci 8980: Data Mining (Fall 2002) Vipin Kumar Army High Performance Computing Research Center Department of Computer.
Research Project Mining Negative Rules in Large Databases using GRD.
Mining Sequences. Examples of Sequence Web sequence:  {Homepage} {Electronics} {Digital Cameras} {Canon Digital Camera} {Shopping Cart} {Order Confirmation}
1 Fast Algorithms for Mining Association Rules Rakesh Agrawal Ramakrishnan Srikant Slides from Ofer Pasternak.
Bayesian Decision Theory Making Decisions Under uncertainty 1.
Basic Data Mining Techniques
17.5 Rule Learning Given the importance of rule-based systems and the human effort that is required to elicit good rules from experts, it is natural to.
Decision Tree Problems CSE-391: Artificial Intelligence University of Pennsylvania Matt Huenerfauth April 2005.
Mohammad Ali Keyvanrad
Data Mining Association Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 6 Introduction to Data Mining By Tan, Steinbach, Kumar Lecture.
Modul 7: Association Analysis. 2 Association Rule Mining  Given a set of transactions, find rules that will predict the occurrence of an item based on.
1 Bayesian Methods. 2 Naïve Bayes New data point to classify: X=(x 1,x 2,…x m ) Strategy: – Calculate P(C i /X) for each class C i. – Select C i for which.
Bab 5 Classification: Alternative Techniques Part 1 Rule-Based Classifer.
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
Outline Knowledge discovery in databases. Data warehousing. Data mining. Different types of data mining. The Apriori algorithm for generating association.
CSE4334/5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai.
Fast Algorithms for Mining Association Rules Rakesh Agrawal and Ramakrishnan Srikant VLDB '94 presented by kurt partridge cse 590db oct 4, 1999.
Association Rule Mining Data Mining and Knowledge Discovery Prof. Carolina Ruiz and Weiyang Lin Department of Computer Science Worcester Polytechnic Institute.
Mining Quantitative Association Rules in Large Relational Tables ACM SIGMOD Conference 1996 Authors: R. Srikant, and R. Agrawal Presented by: Sasi Sekhar.
Association Rule Mining
Decision Trees, Part 1 Reading: Textbook, Chapter 6.
MULTI-INTERVAL DISCRETIZATION OF CONTINUOUS VALUED ATTRIBUTES FOR CLASSIFICATION LEARNING KIRANKUMAR K. TAMBALKAR.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.2: Classification Rules Rodney Nielsen Many.
Association Analysis (3)
Elsayed Hemayed Data Mining Course
Outline Decision tree representation ID3 learning algorithm Entropy, Information gain Issues in decision tree learning 2.
Reducing Number of Candidates Apriori principle: – If an itemset is frequent, then all of its subsets must also be frequent Apriori principle holds due.
CS4445 Data Mining B term WPI Solutions HW4: Classification Rules using RIPPER By Chiying Wang 1.
1 Data Mining Lecture 6: Association Analysis. 2 Association Rule Mining l Given a set of transactions, find rules that will predict the occurrence of.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
Chapter 3 Data Mining: Classification & Association Chapter 4 in the text box Section: 4.3 (4.3.1),
Reducing Number of Candidates
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Data Science Algorithms: The Basic Methods
Frequent Pattern Mining
Ordering of Hypothesis Space
Association Rule Mining
DIRECT HASHING AND PRUNING (DHP) ALGORITHM
COMP5331 FP-Tree Prepared by Raymond Wong Presented by Raymond Wong
Data Mining CSCI 307, Spring 2019 Lecture 18
Presentation transcript:

CS4445/B12 Provided by: Kenneth J. Loomis

CLASSIFICATION RULES: RIPPER ALGORITHM

The first thing that needs to be determined is the consequence of the rule: Recall that a rule is made up of an antecedent  consequence. The table below contains the frequency counts of the possible consequences of the rules from the userprofile dataset using budget as the classification attribute: RuleFrequency …  budget=low35 …  budget=medium91 …  budget=high5 …  budget=?7 We can see that budget=high has the lowest frequency count in our training dataset, so we choose that as the first antecedent that we will find rules for. Note: I have included missing values here as one could classify the target as missing. Alternately, these instances could be removed.

Next we attempt to find the first condition in the antecedent. We need only look at possible conditions that exists in the 5 instances that have budget=high. The list of possible conditions are in the table below. Rule: ___ -> budget=high smoker=trueambience=familypersonality=hard-worker smoker=falseambience=friendspersonality=conformist drink_level=abstemioustransport=car ownerpersonality=hunter-ostentatious drink_level=casual drinkertransport=publicpersonality=thrifty-protector drink_level=social drinkermarital_status=singlereligion=none dress_preference=no preferenceinterest=technologyreligion=mormon dress_preference=informalinterest=nonereligion=christian dress_preference=formalinterest=varietyactivity=student

Here we see a list of the information gain for each of the possible first condition in the antecedent Rule: ___ -> budget=highInfo GainRule: ___ -> budget=highInfo Gain smoker=true0.0862marital_status=single smoker=false interest=technology drink_level=abstemious2.0974interest=none drink_level=casual drinker interest=variety drink_level=social drinker personality=hard-worker dress_preference=no preference0.1174personality=conformist dress_preference=informal personality=hunter-ostentatious dress_preference=formal personality=thrifty-protector ambience=family religion=none ambience=friends2.5440religion=mormon transport=car owner6.7865religion=christian transport=public activity=student

Next we attempt to find the second condition in the antecedent. We need only look at possible conditions that exists in the 4 instances that have transport = car owner and budget=high. The list of possible conditions are in the table below. Rule: transport=car owner and ___ -> budget=high smoker=falseambience=friendspersonality=thrifty-protector drink_level=abstemiousmarital_status=singlereligion=none drink_level=casual drinkerinterest=technologyreligion=mormon dress_preference=no preferenceinterest=nonereligion=christian dress_preference=informalinterest=varietyactivity=student dress_preference=elegantpersonality=hard-worker ambience=familypersonality=hunter-ostentatious

Here we see a list of the information gain for each of the possible second condition in the antecedent Rule: transport=car owner and ___ -> budget=high Info GainRule: transport=car owner and ___ -> budget=high Info Gain smoker=false2.5121interest=none drink_level=abstemious5.0173interest=variety drink_level=casual drinker personality=hard-worker dress_preference=no preference personality=hunter-ostentatious dress_preference=informal0.7655personality=thrifty-protector dress_preference=elegant3.0875religion=none ambience=family religion=mormon ambience=friends1.5075religion=christian marital_status=single2.7570activity=student interest=technology2.5602

Next we attempt to find the third condition in the antecedent. We need only look at possible conditions that exists in the 3 instances that have transport = car owner and drink_level = abstemious and budget=high. The list of possible conditions are in the table below. Rule: transport=car owner and drink_level=abstemious and ___ -> budget=high smoker=falseinterest=technologypersonality=thrifty-protector dress_preference=no preferenceinterest=nonereligion=none dress_preference=formalinterest=varietyreligion=catholic ambience=familypersonality=hard-workerreligion=christian ambience=friendspersonality=hunter-ostentatiousactivity=student marital_status=single

Here we see a list of the information gain for each of the possible third conditions in the antecedent Rule: transport=car owner and drink_level=abstemious and ___ -> budget=high Info GainRule: transport=car owner and drink_level=abstemious and ___ -> budget=high Info Gain smoker=false0interest=variety dress_preference=no preference personality=hard-worker dress_preference=formal1.4513personality=hunter-ostentatious ambience=family personality=thrifty-protector ambience=friends2.8300religion=none marital_status=single1.2415religion=catholic interest=technology0.4515religion=christian interest=none activity=student.01826

Since the following rule results in the highest information gain we select that as the third condition of our rule: transport = car owner and drink_level = abstemious and ambience = friends  budget = high: Note that this rule covers only positive examples (i.e., budget=high data instances). Since it doesn’t cover negative examples, then there is no need to add more conditions to the rule. RIPPER’s construction of the first rule is now complete.

First rule: transport = car owner and drink_level = abstemious and ambience = friends  budget = high: In order to decide if/how to prune this rule, RIPPER will: use a validation set (that is, a piece of the training set that was kept apart and not used to construct the rule) use a metric for pruning: v = (p-n)/(p+n) where p: # of positive examples covered by the rule in the validation set n: # of negative examples covered by the rule in the validation set pruning method: deletes any final sequence of conditions that maximizes v. That is, it calculates v for each of the following pruned versions of the rule and keeps the version of the rule with maximum v: transport = car owner & drink_level = abstemious & ambience = friends  budget = high transport = car owner & drink_level = abstemious  budget = high transport = car owner  budget = high  budget = high

ASSOCIATION RULES: APRIORI ALGORITHM

We begin the Apriori algorithm by determining the order: Here I will use the order that the attributes appear and the values for each attribute in alphabetical order. Then all the possible single item rules are generated and the support calculated for each rule. The following slide shows the complete list of possible items in the rule. Support is calculated in the following manner: Since we know the minimum acceptable support count is 55, we need only look at the numerator of this ratio to determine whether or not to keep this item.

Candidate Itemsets with Support Count smoker=false109transport=on foot14religion=christian7 smoker=true26transport=public82religion=jewish1 drink_level=abstemious51marital_status=single122religion=mormon1 drink_level=casual drinker47marital_status=married10religion=none30 drink_level=social drinker40interest=eco-friendly16activity=professional15 dress_preference=elegant4interest=none30activity=student113 dress_preference=formal41interest=technology36activity=unemployed2 dress_preference=informal53interest=variety50activity=working-class1 dress_preference=no preference35personality=conformist7budget=high5 ambience=family70personality=hard-worker61budget-low35 ambience=friends46personality=hunter-ostentatious12budget=medium91 ambience=solitary16personality=thrifty-protector58 transport=car owner34religion=catholic99 We keep the ones in bold as they meet the minimum support threshold.

Itemsets with Support smoker=false109 ambience=family70 transport=public82 marital_status=single122 personality=hard-worker61 personality=thrifty-protector58 religion=catholic99 activity=student113 budget=medium91 We keep the following item sets as they contain enough support, and use these item sets to generate candidate item sets for the next level.

We merge pairs from the level 1 set. Since there are no prefixes here then we must consider all combinations. (Continued on next slide) Candidate Itemsets with Support Count smoker=false, ambience=family 59 smoker=false, budget=medium 75 ambience=family, budget=medium 54 smoker=false, transport=public 69 ambience=family, transport=public 46 transport=public, marital_status=single 76 smoker=false, marital_status=single 98 ambience=family, marital_status=single 63 transport=public, personality=hard-worker 28 smoker=false, personality=hard-worker 49 ambience=family, personality=hard-worker 26 transport=public, personality=thrifty-protector 44 smoker=false, personality=thrifty-protector 48 ambience=family, personality=thrifty-protector 33 transport=public, religion=catholic 62 smoker=false, religion=catholic 79 ambience=family, religion=catholic 57 transport=public, activity=student 71 smoker=false, activity=student 90 ambience=family, activity=student 61 transport=public, budget=medium 54

Candidate Itemsets with Support Count marital_status=single, personality=hard-worker 52 personality=hard-worker budget=medium 40 marital_status=single, personality=thrifty-protector 51 personality=thrifty-protector, religion=catholic 45 marital_status=single, religion=catholic 91 personality=thrifty-protector, activity=student 50 marital_status=single, activity=student 107 personality=thrifty-protector, budget=medium 41 marital_status=single, budget=medium 79 religion=catholic, activity=student 84 personality=hard-worker, personality=thrifty-protector 0 religion=catholic, budget=medium 67 personality=hard-worker, religion=catholic 40 activity=student, budget=medium 71 personality=hard-worker, activity=student 46

Itemsets with Support Count smoker=false, ambience=family 59 ambience=family, marital_status=single 63 marital_status=single, religion=catholic 91 smoker=false, transport=public 69 ambience=family, religion=catholic 57 marital_status=single, activity=student 107 smoker=false, marital_status=single 98 ambience=family, activity=student 61 marital_status=single, budget=medium 79 smoker=false, religion=catholic 79 transport=public, marital_status=single 76 religion=catholic, activity=student 84 smoker=false, activity=student 90 transport=public, religion=catholic 62 religion=catholic, budget=medium 67 smoker=false, budget=medium 75 transport=public, activity=student 71 activity=student, budget=medium 71 We keep the following item sets as they contain enough support, and use these item sets to generate candidate item sets for the next level.

We generate the next level of candidate sets, but before we calculate the support we can use the Apriori principle to determine if they are viable candidates. Itemsets from Level 2 smoker=false, ambience=family ambience=family, marital_status=single marital_status=single, religion=catholic smoker=false, transport=public ambience=family, religion=catholic marital_status=single, activity=student smoker=false, marital_status=single ambience=family, activity=student marital_status=single, budget=medium smoker=false, religion=catholic transport=public, marital_status=single religion=catholic, activity=student smoker=false, activity=student transport=public, religion=catholic religion=catholic, budget=medium smoker=false, budget=medium transport=public, activity=student activity=student, budget=medium

First we determine the candidates by “joining” itemsets with like prefixes. (i.e., the first k-1 items in the items sets are the same) Here we need only match the first item in the itemset. Itemsets from Level 2 smoker=false, ambience=family ambience=family, marital_status=single marital_status=single, religion=catholic smoker=false, transport=public ambience=family, religion=catholic marital_status=single, activity=student smoker=false, marital_status=single ambience=family, activity=student marital_status=single, budget=medium smoker=false, religion=catholic transport=public, marital_status=single religion=catholic, activity=student smoker=false, activity=student transport=public, religion=catholic religion=catholic, budget=medium smoker=false, budget=medium transport=public, activity=student activity=student, budget=medium

That results in this set of potential candidate itemsets. Potential Candidate Itemsets smoker=false, ambience=family, transport=public smoker=false, transport=public, religion=catholic smoker=false, activity=student, budget=medium transport=public, religion=catholic, activity=student smoker=false, ambience=family, marital_status=single smoker=false, transport=public, activity=student ambience=family, marital_status=single, religion=catholic marital_status=single, religion=catholic, activity=student smoker=false, ambience=family, religion=catholic smoker=false, transport=public, budget=medium ambience=family, marital_status=single, activity=student marital_status=single, religion=catholic, budget=medium smoker=false, ambience=family, activity=student smoker=false, marital_status=single, religion=catholic ambience=family, religion=catholic, activity=student marital_status=single, activity=student, budget=medium smoker=false, ambience=family, budget=medium smoker=false, marital_status=single, activity=student transport=public, marital_status=single, religion=catholic religion=catholic, activity=student, budget=medium smoker=false, transport=public, marital_status=single smoker=false, marital_status=single, budget=medium transport=public, marital_status=single, activity=student

We have one final step before calculating the support: we can eliminate unnecessary candidates. We must check that all subsets of size 2 in each of these itemsets also existed in the level 2 set. We can make this a little easier by ignoring the prefix subsets as we know those existed because we used them to create the itemsets. The following itemsets can be removed as the bolded subsets do not appear in the Level 2 itemsets. This leaves us the candidate itemsets on the next slide. Candidate Itemsets That Can be Removed smoker=false, ambience=family, transport=public smoker=false, ambience=family, budget=medium smoker=false, transport=public, budget=medium

Finally we can calculate the support for these candidate itemsets. Candidate Itemsets with Support Count smoker=false, ambience=family, marital_status=single 53 smoker=false, transport=public, activity=student 58 ambience=family, marital_status=single, religion=catholic 50 transport=public, religion=catholic, activity=student 59 smoker=false, ambience=family, religion=catholic 46 smoker=false, marital_status=single, religion=catholic 72 ambience=family, marital_status=single, activity=student 57 marital_status=single, religion=catholic, activity=student 80 smoker=false, ambience=family, activity=student 52 smoker=false, marital_status=single, activity=student 85 ambience=family, religion=catholic, activity=student 51 marital_status=single, religion=catholic, budget=medium 80 smoker=false, transport=public, marital_status=single 63 smoker=false, marital_status=single, budget=medium 65 transport=public, marital_status=single, religion=catholic 57 marital_status=single, activity=student, budget=medium 59 smoker=false, transport=public, religion=catholic 52 smoker=false, activity=student, budget=medium 58 transport=public, marital_status=single, activity=student 67 religion=catholic, activity=student, budget=medium 53

Level 3 Itemsets with Support smoker=false, transport=public, marital_status=single 63 smoker=false, activity=student, budget=medium 58 marital_status=single, religion=catholic, activity=student 80 smoker=false, transport=public, activity=student 58 ambience=family, marital_status=single, activity=student 57 marital_status=single, religion=catholic, budget=medium 80 smoker=false, marital_status=single, religion=catholic 72 transport=public, marital_status=single, religion=catholic 57 marital_status=single, activity=student, budget=medium 59 smoker=false, marital_status=single, activity=student 85 transport=public, marital_status=single, activity=student 67 smoker=false, marital_status=single, budget=medium 65 transport=public, religion=catholic, activity=student 59 We keep the following item sets as they contain enough support, and use these item sets to generate candidate item sets for the next level.

We generate the next level of candidate sets, but before we calculate the support we can use the Apriori principle to determine if they are viable candidates. Level 3 Itemsets smoker=false, transport=public, marital_status=single smoker=false, activity=student, budget=medium marital_status=single, religion=catholic, activity=student smoker=false, transport=public, activity=student ambience=family, marital_status=single, activity=student marital_status=single, religion=catholic, budget=medium smoker=false, marital_status=single, religion=catholic transport=public, marital_status=single, religion=catholic marital_status=single, activity=student, budget=medium smoker=false, marital_status=single, activity=student transport=public, marital_status=single, activity=student smoker=false, marital_status=single, budget=medium transport=public, religion=catholic, activity=student

Finally we can calculate the support for these candidate itemsets. We generate the next level of candidate sets, but before we calculate the support we can use the Apriori principle to determine if they are viable candidates. Level 3 Itemsets smoker=false, transport=public, marital_status=single smoker=false, activity=student, budget=medium marital_status=single, religion=catholic, activity=student smoker=false, transport=public, activity=student ambience=family, marital_status=single, activity=student marital_status=single, religion=catholic, budget=medium smoker=false, marital_status=single, religion=catholic transport=public, marital_status=single, religion=catholic marital_status=single, activity=student, budget=medium smoker=false, marital_status=single, activity=student transport=public, marital_status=single, activity=student smoker=false, marital_status=single, budget=medium transport=public, religion=catholic, activity=student First we determine the candidates by “joining” itemsets with like prefixes. (i.e., the first k-1 items in the items sets match) Here we need only match the first two items in the itemset.

Potential Candidate Item Sets smoker=false, transport=public, marital_status=single, activity=student smoker=false, marital_status=single, activity=student, budget=medium smoker=false, marital_status=single, religion=catholic, activity=student transport=public, marital_status=single, religion=catholic, activity=student smoker=false, marital_status=single, religion=catholic, budget=medium marital_status=single, religion=catholic, activity=student, budget=medium That results in this set of candidate itemsets. We have one final step before calculating the support: we can eliminate unnecessary candidates. We must check that all subsets of size 3 in each of these itemsets also existed in the level 3 set. We can make this a little easier by ignoring the prefix subsets as we know those existed because we used them to create the itemsets. Here we again eliminate candidates from consideration, the offending subsets are bolded.

Candidate Itemsets with Support Count smoker=false, marital_status=single, religion=catholic, activity=student 63 smoker=false, marital_status=single, activity=student, budget=medium 53 In the end we keep only one single itemset that has enough support for this level. The following slide depicts the complete itemset. Level 4 Itemsets with Support Count smoker=false, marital_status=single, religion=catholic, activity=student 63

Itemsets with Support Count smoker=false109 smoker=false, marital_status=single 98 marital_status=single, religion=catholic 91 smoker=false, marital_status=single, budget=medium 65 ambience=family70 smoker=false, religion=catholic 79 marital_status=single, activity=student 107 smoker=false, activity=student, budget=medium 58 marital_status=single122 smoker=false, activity=student 90 marital_status=single, budget=medium 79 ambience=family, marital_status=single, activity=student 57 personality=hard-worker61 smoker=false, budget=medium 75 religion=catholic, activity=student 84 transport=public, marital_status=single, religion=catholic 57 transport=public82 ambience=family, marital_status=single 63 religion=catholic, budget=medium 67 transport=public, marital_status=single, activity=student 67 religion=catholic99 ambience=family, religion=catholic 57 activity=student, budget=medium 71 transport=public, religion=catholic, activity=student 59 activity=student113 ambience=family, activity=student 61 smoker=false, transport=public, marital_status=single 63 marital_status=single, religion=catholic, activity=student 80 budget=medium91 transport=public, marital_status=single 76 smoker=false, transport=public, activity=student 58 marital_status=single, religion=catholic, budget=medium 80 smoker=false, ambience=family 59 transport=public, religion=catholic 62 smoker=false, marital_status=single, religion=catholic 72 marital_status=single, activity=student, budget=medium 59 smoker=false, transport=public 69 transport=public, activity=student 71 smoker=false, marital_status=single, activity=student 85 smoker=false, marital_status=single, religion=catholic, activity=student 63

Largest itemset: Let’s call this itemset I4: I4: smoker=false, marital_status=single, religion=catholic, activity=student Rules constructed from I4 with 2 items in the antecedent:  R1: smoker=false, marital_status=single  religion=catholic, activity=student conf(R1) = supp(I4)/supp(smoker=false, marital_status=single ) = 63/ 98 = 64.28%  R2: smoker=false, religion=catholic  marital_status=single, activity=student conf(R2) = supp(I4)/supp(smoker=false, religion=catholic ) = 63/ 79 = 79.74%  R3: smoker=false, activity=student  marital_status=single, religion=catholic conf(R3) = supp(I4)/supp(smoker=false, activity=student ) = 63/ 90= 70%  R4: marital_status=single, religion=catholic  smoker=false, activity=student conf(R4) = supp(I4)/supp(marital_status=single, religion=catholic ) = 63/ 91 = 69.23%  R5: marital_status=single, activity=student  smoker=false, religion=catholic conf(R5) = supp(I4)/supp(marital_status=single, activity=student ) = 63/ 107 = 58.87%  R6: religion=catholic, activity=student  smoker=false, marital_status=single conf(R6) = supp(I4)/supp(religion=catholic, activity=student) = 63/ 84 = 75%