Presentation is loading. Please wait.

Presentation is loading. Please wait.

SAD: 6º Projecto.

Similar presentations


Presentation on theme: "SAD: 6º Projecto."— Presentation transcript:

1 SAD: 6º Projecto

2

3

4 Lift Charts Comparing classifiers: 1,000,000 prospective respondents
prediction that 0.1% of all households (1,000,000) will respond prediction that 0.4% of a specified 100,000 homes will respond. lift factor=increase in response rate=4 Given a classifier that outputs probabilities for the predicted class value for each test instance, what to do?

5 Lift Factor sample success proportion lift factor
= (number of positive instances in sample) / sample size lift factor = (sample success proportion) / (total test set success proportion)

6 Lift

7 Evaluation: The confusion matrix
Incorrectly classified instances a b  classified as | a = yes | b = no Correctly classified instances Comments: For a boolean classification, the entropy is 0 if all entities belong to the same class; the entropy is 1 if the collection contains an equal number of positive and negative examples. Typical measure of entropy: bits of information needed for encoding the classification. Note that the first term for gain is the entropy of the original collection, and the second term is the expected value of the entropy after C is partitioned using attribute A. Gain is the exptected reduction in entropy caused by knowing the value of attribute A.

8

9 b) (1 pts) Perform Cross Validation of all your algorithms with Fold Count 4, 8, Maximum Cases should be Which algorithm is the best, which varies less? Which is the better choice?

10 Decision Tree

11

12 Naïve Bayes

13

14 NN

15

16 Paired Sample t Test Given a set of paired observations
(from two normal populations) A B =A-B x1 y1 x1-x2 x2 y2 x2-y2 x3 y3 x3-y3 x4 y4 x4-y4 x5 y5 x5-y5

17 Calculate the mean and the standard deviation s of the the differences 
H0: =0 (no difference) H0: =k (difference is a constant)

18 DT

19 NB

20 NN

21 DT [182 – 188 – ] NB [184 – – ] NN [163.1 – 166 – 168]

22 DT [89.28– 94 – 98.72] NB [91.73 – – ] NN [76.38– –90.38]

23 Shannon formalized these intuitions
Given a universe of messages M={m1,m2,...,mn} and a probability p(mi) for the occurrence of each message, the information content (also called entropy)of a message M is given

24 The amount of information needed to complete the tree is defined as weighted average of the information content of each sub tree by the percentage of the examples present C a set of training instances. If property (for example income) with n values, C will be divided into the subsets {C1,C2,..,Cn} Expected information needed to complete the tree after making P root

25 The gain from the property P is computed by subtracting the expected information to complete E(P) fro the total information

26 2. (6pts) Decision Tree

27

28

29


Download ppt "SAD: 6º Projecto."

Similar presentations


Ads by Google