PROJECTS ON SUPERVISED AND FACTORIZATION BAYESIAN NETWORKS Course 2007/2008 Concha Bielza, Pedro Larrañaga Universidad Politécnica de Madrid.

Slides:



Advertisements
Similar presentations
Active Learning based on Bayesian Networks Luis M. de Campos, Silvia Acid and Moisés Fernández.
Advertisements

Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
S3-SEMINAR ON DATA MINING -BAYESIAN NETWORKS- B. INFERENCE Master Universitario en Inteligencia Artificial Concha Bielza, Pedro Larrañaga Computational.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Weka. Preprocessing Opening a file Editing a file Visualize a variable.
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
Weka. Preprocessing Opening a file Editing a file Visualize a variable.
Rafi Bojmel supervised by Dr. Boaz Lerner Automatic Threshold Selection for conditional independence tests in learning a Bayesian network.
Bayesian Network Classifiers for Identifying the Slope of the customer Lifecycle of Long-Life Customers Authored by: Bart Baesens, Geert Vertraeten, Dirk.
Overview Full Bayesian Learning MAP learning
1 Augmented Bayesian Classifier Beijie Xu. 2 problem to be solved NAÏVE Bayesian Classifier –Why naïve? - conditional independence among attributes –Is.
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Lecture 16: Wrap-Up COMP 538 Introduction of Bayesian networks.
PROJECTS ON INTRODUCTION AND INFERENCE Course 2007/2008 Concha Bielza, Pedro Larrañaga Universidad Politécnica de Madrid.
Evolution of descent directions Alejandro Sierra Escuela Politécnica Superior Universidad Autónoma de Madrid Iván Santibáñez Koref Bionik und Evolutionstechnik.
PGM: Tirgul 11 Na?ve Bayesian Classifier + Tree Augmented Na?ve Bayes (adapted from tutorial by Nir Friedman and Moises Goldszmidt.
Practice of Bayesian Networks
Bayesian networks practice. Semantics e.g., P(j  m  a   b   e) = P(j | a) P(m | a) P(a |  b,  e) P(  b) P(  e) = … Suppose we have the variables.
Data Mining: A Closer Look Chapter Data Mining Strategies (p35) Moh!
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes April 3, 2012.
An Exercise in Machine Learning
THE HONG KONG UNIVERSITY OF SCIENCE & TECHNOLOGY CSIT 5220: Reasoning and Decision under Uncertainty L10: Model-Based Classification and Clustering Nevin.
Bayesian Networks 4 th, December 2009 Presented by Kwak, Nam-ju The slides are based on, 2nd ed., written by Ian H. Witten & Eibe Frank. Images and Materials.
A Brief Introduction to Graphical Models
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Topics on Final Perceptrons SVMs Precision/Recall/ROC Decision Trees Naive Bayes Bayesian networks Adaboost Genetic algorithms Q learning Not on the final:
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
1 1 Slide Evaluation. 2 2 n Interactive decision tree construction Load segmentchallenge.arff; look at dataset Load segmentchallenge.arff; look at dataset.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Bayesian Networks Martin Bachler MLA - VO
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
Aprendizagem Computacional Gladys Castillo, UA Bayesian Networks Classifiers Gladys Castillo University of Aveiro.
Hands-on predictive models and machine learning for software Foutse Khomh, Queen’s University Segla Kpodjedo, École Polytechnique de Montreal PASED - Canadian.
Bayesian Learning Chapter Some material adapted from lecture notes by Lise Getoor and Ron Parr.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Statistical Inference (By Michael Jordon) l Bayesian perspective –conditional perspective—inferences.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Active learning Haidong Shi, Nanyi Zeng Nov,12,2008.
Artificial Intelligence Project #3 : Diagnosis Using Bayesian Networks May 19, 2005.
An Exercise in Machine Learning
Rotem Golan Department of Computer Science Ben-Gurion University of the Negev, Israel.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Dimensionality Reduction in Unsupervised Learning of Conditional Gaussian Networks Authors: Pegna, J.M., Lozano, J.A., Larragnaga, P., and Inza, I. In.
***Classification Model*** Hosam Al-Samarraie, PhD. CITM-USM.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
Machine Learning in Practice Lecture 21 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss Pedro Domingos, Michael Pazzani Presented by Lu Ren Oct. 1, 2007.
Bias Management in Time Changing Data Streams We assume data is generated randomly according to a stationary distribution. Data comes in the form of streams.
Genetic Algorithms Schematic of neural network application to identify metabolites by mass spectrometry (MS) Developed by Dr. Lars Kangas Input to Genetic.
Data Science Credibility: Evaluating What’s Been Learned
Latent variable discovery in classification models
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Bayesian Classification
Data Mining Lecture 11.
Prepared by: Mahmoud Rafeek Al-Farra
Adapted from: Prof. Pedro Larrañaga Technical University of Madrid
CSE P573 Applications of Artificial Intelligence Bayesian Learning
Opening Weka Select Weka from Start Menu Select Explorer Fall 2003
CSCI N317 Computation for Scientific Applications Unit Weka
Sofia Pediaditaki and Mahesh Marina University of Edinburgh
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
Practice Project Overview
Data Mining CSCI 307, Spring 2019 Lecture 8
Presentation transcript:

PROJECTS ON SUPERVISED AND FACTORIZATION BAYESIAN NETWORKS Course 2007/2008 Concha Bielza, Pedro Larrañaga Universidad Politécnica de Madrid

Hugin Lite Build a Bayesian network of your invention with six nodes and binary variables How to Build BNs (Hugin GUI Help) 2. Generate 50, 100, 200 and 400 cases from the previously built Bayesian network Case Generator (Hugin GUI Help) 3. Structure learning with PC and NPC algorithms with two level of significance (0.05 and 0.10) Structure Learning (Hugin GUI Help) 4. Parameter learning with the EM learning algorithm EM learning (Hugin GUI Help) Factorization Exercise

Hugin Lite PC NPC Simulation Simulation Simulation Simulation Hamming distance between the structure of the original Bayesian network, and the one obtained after learning

BAYESIA Generate two data bases (50 and 500 instances and different percentage of missing data) from the “Asia.xbl” Bayesian network 2. Apply the following learning algoithms: “EQ”, “SopLEQ”, “Tabo” and “TaboOrder” to both data bases 3. Compare the induced Bayesian networks with the “Asia.xbl” 4. Obtain information in Internet about the learning algorithms Factorization Exercice

Weka 1.Using the “tips-discrete-cfs9.arff” dataset 2. Learn Bayesian network structures with: - Conditional independence tests - Local search - Global search 3. Estimate the parameters: - Simple estimation - BMA estimator Factorization Exercice

BAYESIA Supervised Exercice 1.Generate 3 files (100, 200 and 400 cases) from the “Asia.xbl” Bayesian network 2. Choose variable “Cancer” as the class (target) variable 3. Induce the following classifiers: Naive Bayes Augmented naive Bayes Markov blanket 4. Compare the accuracies of the different models in the 3 files

Weka 1.Open the file “tips-discrete-cfs9.arff” 2. Learn naive Bayes and TAN models 3. Obtain the corresponding accuracies with a 10-fold cv validation method 4. Repeat the exercice with a FSS method (Select Attributes in Weka) Supervised Exercice