Download presentation
Presentation is loading. Please wait.
Published byCleopatra Gallagher Modified over 9 years ago
1
Entropy-based & ChiMerge Data Discretization Feb. 12, 2008 Team #4: Seunghyun Kim Craig Dunham Suryo Muljono Albert Lee
2
Entropy-based discretization Table 6.1 Class-labeled training tuples from the AllElectronics customer database (page 299). RIDageincomeStudentCredit_ratingClass: buy_computer 1YouthHighNoFaireNo 2YouthHighNoExcellentNo 3Middle_ageedHighNoFaireYes 4SeniorMediumNoFaireYes 5SeniorLowYesFaireYes 6SeniorLowYesExcellentNo 7Middle_agedLowYesExcellentYes 8YouthMediumNoFaireNo 9YouthLowYesFaireYes 10SeniorMediumYesFaireYes 11YouthMediumYesExcellentYes 12Middle_ageedMediumNoExcellentYes 13Middle_ageedHighYesFaireYes 14SeniorMediumNoExcellentNo
3
Entropy-based (Cont’d) Information gain Info(D) = = 0.940 bits Info age (D) = = 0.649 bits
4
Entropy-based (Cont’d) Gain(A) = Info(D) – Info A (D). Gain(age) = Info(D) – Info age (D) = 0.940 – 0.694 = 0.246 bits Gain(income)= Info(D) – Info income (D) = 0.940 – 0.911 = 0.029 bits Gain(student)= Info(D) – Info student (D)= 0.940 – 0.694 = 0.152 bits Gain(credit) = Info(D) – Info credit (D) = 0.940 – 0.892 = 0.04 bits
5
Entropy-based (Cont’d) AllElectronics customer database Age ? SeniorMiddle_ageYouth
6
Entropy-based (Cont’d) AllElectronics customer database Age ? SeniorMiddle Youth Student? Credit? Student Non Student ExcellentFair yes no
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.