Cost-Sensitive Bayesian Network algorithm Introduction: Machine learning algorithms are becoming an increasingly important area for research and application.

Slides:



Advertisements
Similar presentations
A Fully Distributed Framework for Cost-sensitive Data Mining Wei Fan, Haixun Wang, and Philip S. Yu IBM T.J.Watson, Hawthorne, New York Salvatore J. Stolfo.
Advertisements

Decision Tree Evolution using Limited number of Labeled Data Items from Drifting Data Streams Wei Fan 1, Yi-an Huang 2, and Philip S. Yu 1 1 IBM T.J.Watson.
On Comparing Classifiers: Pitfalls to Avoid and a Recommended Approach Author: Steven L. Salzberg Presented by: Zheng Liu.
Machine Learning Instance Based Learning & Case Based Reasoning Exercise Solutions.
Chapter 4 Pattern Recognition Concepts: Introduction & ROC Analysis.
Imbalanced data David Kauchak CS 451 – Fall 2013.
Decision Tree Approach in Data Mining
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
1 Test-Cost Sensitive Naïve Bayes Classification X. Chai, L. Deng, Q. Yang Dept. of Computer Science The Hong Kong University of Science and Technology.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Learning on Probabilistic Labels Peng Peng, Raymond Chi-wing Wong, Philip S. Yu CSE, HKUST 1.
1. Abstract 2 Introduction Related Work Conclusion References.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Presented by Zeehasham Rasheed
Boosting Main idea: train classifiers (e.g. decision trees) in a sequence. a new classifier should focus on those cases which were incorrectly classified.
Ordinal Decision Trees Qinghua Hu Harbin Institute of Technology
Introduction to machine learning
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
Enterprise systems infrastructure and architecture DT211 4
InCob A particle swarm based hybrid system for imbalanced medical data sampling Pengyi Yang School of Information Technologies.
Evaluating Classifiers
1 © Goharian & Grossman 2003 Introduction to Data Mining (CS 422) Fall 2010.
CS490D: Introduction to Data Mining Prof. Chris Clifton April 14, 2004 Fraud and Misuse Detection.
Learning from Imbalanced, Only Positive and Unlabeled Data Yetian Chen
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Using Neural Networks in Database Mining Tino Jimenez CS157B MW 9-10:15 February 19, 2009.
Human Gesture Recognition Using Kinect Camera Presented by Carolina Vettorazzo and Diego Santo Orasa Patsadu, Chakarida Nukoolkit and Bunthit Watanapa.
Classification and Prediction (cont.) Pertemuan 10 Matakuliah: M0614 / Data Mining & OLAP Tahun : Feb
Page 1 Ming Ji Department of Computer Science University of Illinois at Urbana-Champaign.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A data mining approach to the prediction of corporate failure.
An Overview of Intrusion Detection Using Soft Computing Archana Sapkota Palden Lama CS591 Fall 2009.
Comparison of Bayesian Neural Networks with TMVA classifiers Richa Sharma, Vipin Bhatnagar Panjab University, Chandigarh India-CMS March, 2009 Meeting,
1 CS 391L: Machine Learning: Experimental Evaluation Raymond J. Mooney University of Texas at Austin.
1 KDD-09, Paris France Quantification and Semi-Supervised Classification Methods for Handling Changes in Class Distribution Jack Chongjie Xue † Gary M.
Computational Intelligence: Methods and Applications Lecture 12 Bayesian decisions: foundation of learning Włodzisław Duch Dept. of Informatics, UMK Google:
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
Classification Techniques: Bayesian Classification
1 Pattern Recognition Pattern recognition is: 1. A research area in which patterns in data are found, recognized, discovered, …whatever. 2. A catchall.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Classification And Bayesian Learning
October 2-3, 2015, İSTANBUL Boğaziçi University Prof.Dr. M.Erdal Balaban Istanbul University Faculty of Business Administration Avcılar, Istanbul - TURKEY.
GENDER AND AGE RECOGNITION FOR VIDEO ANALYTICS SOLUTION PRESENTED BY: SUBHASH REDDY JOLAPURAM.
يادگيري ماشين Machine Learning Lecturer: A. Rabiee
Class Imbalance in Text Classification
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Cost- sensitive boosting for classification of imbalanced.
A Brief Introduction and Issues on the Classification Problem Jin Mao Postdoc, School of Information, University of Arizona Sept 18, 2015.
Combining multiple learners Usman Roshan. Decision tree From Alpaydin, 2010.
1 Discriminative Frequent Pattern Analysis for Effective Classification Presenter: Han Liang COURSE PRESENTATION:
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Lessons Learned from Applications of Machine Learning Robert C. Holte University of Alberta.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Rule Induction for Classification Using
Can-CSC-GBE: Developing Cost-sensitive Classifier with Gentleboost Ensemble for breast cancer classification using protein amino acids and imbalanced data.
Trees, bagging, boosting, and stacking
Dipartimento di Ingegneria «Enzo Ferrari»,
Data Mining Classification: Alternative Techniques
What is Pattern Recognition?
An Inteligent System to Diabetes Prediction
Classification of class-imbalanced data
iSRD Spam Review Detection with Imbalanced Data Distributions
CSCI N317 Computation for Scientific Applications Unit Weka
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Somi Jacob and Christian Bach
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
Presentation transcript:

Cost-Sensitive Bayesian Network algorithm Introduction: Machine learning algorithms are becoming an increasingly important area for research and application in the field of Artificial Intelligence and data mining. One of the most important algorithm is Bayesian network, this algorithm have been widely used in real world applications like medical diagnosis, image recognition, fraud detection, and inference problems. In all of these applications, evaluation method as accuracy is not enough because there are costs involve each decision. For example, in a fraud detection application to predict new case, there are several costs involved when the classifier predicts a fraudulent case as a non-fraudulent case. Also, fraud databases have an unbalanced class distribution which is known to affect learning algorithms adversely. Therefore, this project develops new algorithm that aims to minimize the costs of prediction, misclassification, imbalance data, time and test. In this work, we attempt to create a new cost-sensitive Bayesian network learning algorithm by adapting Bayesian network algorithm, which focuses on accuracy only. There are several ways of adapting our algorithm and make it cost-sensitive, this includes: changing distribution of the data; changing the construction process and even adopting alternative measure in the algorithms that take account of cost; and using Genetic Algorithm to learn structure of BN. This work will apply different approaches such as amending distributions, amending formula, and using Genetic algorithms. Finally, an empirical evaluation of the developed algorithms will be carried on the artificial data sets (e.g diabetes data, lung cancer data, Bank data …etc). Conclusion : In the real world problems such as fraud detection, medical diagnosis, or any decision problem. Often, one class label in dataset such as (Non-fraud class) is very rare and expansive than another class, because the cost of not recognizing some of the instances which belong to the rare class is high. Therefore, most of machine learning methods do not take cost into account. Thus, those algorithms (cost-insensitive algorithms) have a poor result, because ignoring cost might produce a very week model. In reality, misclassification problems (error of classification) are very common problem in real-world data mining when the data is imbalanced in class label. Eman Nashnush University of Salford,Manchester, UK Sponsor in Libya ( Tripoli University ) Hypotheses/The problem Methodology Cost-insensitive Vs. cost-sensitive (Research problem)  A cost-insensitive classifier focus on accuracy only (class label output)..  Cost-sensitive attempt to minimize the expected cost.. Learner Training Data Classifier ($43.45,retail,10040,.. nonfraud) ($246,70,weapon,94583,.,fraud) 1. Decision trees 2. Rules 3. Naive Bayes... Transaction {fraud,nonfraud} Testing data Classifier Class Labels nonfraud fraud ($99.99,pharmacy,10027,...,?) ($1.00,gas,00234,...,?) The previously mentioned problems are happened during classification data set. Therefore, three methods have been proposed to tackle those problems and minimize the expected misclassification cost.  Amend the data distribution to reflect cost.  Amend the formula by modifying the statistical measures to include cost.  Utilize a Genetic algorithm to evolve a 'fittest' Bayesian network. Up to Now, I have investigated experimentally how changing the distribution of data will affect the performance and cost of a Bayesian classifier. I experiment my approach that called “Cost-Sensitive Bayesian Network using Sampling” with 24 data sets from the UCI repository database. I try to compare my proposed algorithm with the existing methods, and also compare the performance of this proposed method with the original algorithm. In the figure below, I show the results of Cost- sensitive Bayes Network algorithm via changing the distributions, and the original Bayes Network algorithm. Results Up to now, two new methods for cost-sensitive Bayesian Network algorithms have been developed and explored: one that uses a black box (Sampling) approach and another that uses a transparent box approach (modifying the statistical measures) that amends the selection measure to take account of costs. The effect of our algorithms are evaluated and compared with other algorithms, such as (MetaCost+J4.8, standard decision tree(J48), and standard Bayesian networks).