Oct 29th, 2001Copyright © 2001, Andrew W. Moore Bayes Net Structure Learning Andrew W. Moore Associate Professor School of Computer Science Carnegie Mellon.

Slides:



Advertisements
Similar presentations
Bayesian network classification using spline-approximated KDE Y. Gurwicz, B. Lerner Journal of Pattern Recognition.
Advertisements

Causal Data Mining Richard Scheines Dept. of Philosophy, Machine Learning, & Human-Computer Interaction Carnegie Mellon.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University Note to other teachers and users of these.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Advanced Methods and Analysis for the Learning and Social Sciences PSY505 Spring term, 2012 February 27, 2012.
Special Topics in Educational Data Mining HUDK5199 Spring term, 2013 March 7, 2013.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Executive Summary: Bayes Nets Andrew W. Moore Carnegie Mellon University, School of Computer Science Note to other teachers.
Collaborative Filtering in iCAMP Max Welling Professor of Computer Science & Statistics.
Assuming normally distributed data! Naïve Bayes Classifier.
An introduction to time series approaches in biosurveillance Professor The Auton Lab School of Computer Science Carnegie Mellon University
COMP 328: Midterm Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Copyright © 2004, Andrew W. Moore Naïve Bayes Classifiers Andrew W. Moore Professor School of Computer Science Carnegie Mellon University
© Daniel S. Weld 1 Naïve Bayes & Expectation Maximization CSE 573.
Learning Bayesian Networks
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Bayes Classifier, Linear Regression 10701/15781 Recitation January 29, 2008 Parts of the slides are from previous years’ recitation and lecture notes,
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes April 3, 2012.
Copyright © 2001, 2004, Andrew W. Moore Clustering with Gaussian Mixtures Andrew W. Moore Professor School of Computer Science Carnegie Mellon University.
Bayesian Networks 4 th, December 2009 Presented by Kwak, Nam-ju The slides are based on, 2nd ed., written by Ian H. Witten & Eibe Frank. Images and Materials.
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Copyright © Andrew W. Moore Slide 1 Gaussians Andrew W. Moore Professor School of Computer Science Carnegie Mellon University
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Copyright © Andrew W. Moore Slide 1 Cross-validation for detecting and preventing overfitting Andrew W. Moore Professor School of Computer Science Carnegie.
Artificial Intelligence Lecture No. 29 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
1 Generative and Discriminative Models Jie Tang Department of Computer Science & Technology Tsinghua University 2012.
1 Support Vector Machines Chapter Nov 23rd, 2001Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines Andrew W. Moore Professor School.
Naïve Bayes William W. Cohen. Probabilistic and Bayesian Analytics Andrew W. Moore School of Computer Science Carnegie Mellon University
Slides for “Data Mining” by I. H. Witten and E. Frank.
Learning In Bayesian Networks. General Learning Problem Set of random variables X = {X 1, X 2, X 3, X 4, …} Training set D = { X (1), X (2), …, X (N)
Machine Learning Lecture 7: SVM Moshe Koppel Slides adapted from Andrew Moore Copyright © 2001, 2003, Andrew W. Moore.
Sep 10th, 2001Copyright © 2001, Andrew W. Moore Learning Gaussian Bayes Classifiers Andrew W. Moore Associate Professor School of Computer Science Carnegie.
Sep 6th, 2001Copyright © 2001, 2004, Andrew W. Moore Learning with Maximum Likelihood Andrew W. Moore Professor School of Computer Science Carnegie Mellon.
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Oct 29th, 2001Copyright © 2001, Andrew W. Moore Eight more Classic Machine Learning algorithms Andrew W. Moore Associate Professor School of Computer Science.
CS 401R: Intro. to Probabilistic Graphical Models Lecture #6: Useful Distributions; Reasoning with Joint Distributions This work is licensed under a Creative.
Nov 30th, 2001Copyright © 2001, Andrew W. Moore PAC-learning Andrew W. Moore Associate Professor School of Computer Science Carnegie Mellon University.
Local Likelihood & other models, Kernel Density Estimation & Classification, Radial Basis Functions.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Revision.
Nov 20th, 2001Copyright © 2001, Andrew W. Moore VC-dimension for characterizing classifiers Andrew W. Moore Associate Professor School of Computer Science.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
Model Comparison.
PDF, Normal Distribution and Linear Regression
Markov Systems, Markov Decision Processes, and Dynamic Programming
Pfizer HTS Machine Learning Algorithms: November 2002
10701 / Machine Learning.
Data Mining Lecture 11.
Model Development Weka User Manual.
Support Vector Machines
CS 188: Artificial Intelligence
Class #212 – Thursday, November 12
CSCE833 Machine Learning Lecture 9 Linear Discriminant Analysis
Support Vector Machines
Linear Discrimination
Analysis of Large Graphs: Overlapping Communities
Presentation transcript:

Oct 29th, 2001Copyright © 2001, Andrew W. Moore Bayes Net Structure Learning Andrew W. Moore Associate Professor School of Computer Science Carnegie Mellon University Note to other teachers and users of these slides. Andrew would be delighted if you found this source material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify them to fit your own needs. PowerPoint originals are available. If you make use of a significant portion of these slides in your own lecture, please include this message, or the following link to the source repository of Andrew’s tutorials: Comments and corrections gratefully received.

Bayes Net Structure: Slide 2 Copyright © 2001, Andrew W. Moore Reminder: A Bayes Net

Bayes Net Structure: Slide 3 Copyright © 2001, Andrew W. Moore Estimating Probability Tables

Bayes Net Structure: Slide 4 Copyright © 2001, Andrew W. Moore Estimating Probability Tables

Bayes Net Structure: Slide 5 Copyright © 2001, Andrew W. Moore Scoring a structure (Which of these fits the data best?) N. Friedman and Z. Yakhini, On the sample complexity of learning Bayesian networks, Proceedings of the 12th conference on Uncertainty in Artificial Intelligence, Morgan Kaufmann, 1996

Bayes Net Structure: Slide 6 Copyright © 2001, Andrew W. Moore Scoring a structure Number of non- redundant parameters defining the net #Records #Attributes Sums over all the rows in the prob- ability table for X j The parent values in the k’th row of X j ’s probability table All these values estimated from data

Bayes Net Structure: Slide 7 Copyright © 2001, Andrew W. Moore Scoring a structure All these values estimated from data This is called a BIC (Bayes Information Criterion) estimate This part is a penalty for too many parameters This part is the training set log- likelihood BIC asymptotically tries to get the structure right. (There’s a lot of heavy emotional debate about whether this is the best scoring criterion)

Bayes Net Structure: Slide 8 Copyright © 2001, Andrew W. Moore Searching for structure with best score

Bayes Net Structure: Slide 9 Copyright © 2001, Andrew W. Moore Learning Methods until today Inputs Classifier Predict category Inputs Density Estimator Prob- ability Inputs Regressor Predict real no. Dec Tree, Sigmoid Perceptron, Sigmoid N.Net, Gauss/Joint BC, Gauss Naïve BC, N.Neigh Joint DE, Naïve DE, Gauss/Joint DE, Gauss Naïve DE Linear Regression, Quadratic Regression, Perceptron, Neural Net, N.Neigh, Kernel, LWR

Bayes Net Structure: Slide 10 Copyright © 2001, Andrew W. Moore Learning Methods added today Inputs Classifier Predict category Inputs Density Estimator Prob- ability Inputs Regressor Predict real no. Dec Tree, Sigmoid Perceptron, Sigmoid N.Net, Gauss/Joint BC, Gauss Naïve BC, N.Neigh Joint DE, Naïve DE, Gauss/Joint DE, Gauss Naïve DE, Bayes Net Structure Learning (Note, can be extended to permit mixed categorical/real values) Linear Regression, Quadratic Regression, Perceptron, Neural Net, N.Neigh, Kernel, LWR

Bayes Net Structure: Slide 11 Copyright © 2001, Andrew W. Moore But also, for free… Inputs Classifier Predict category Inputs Density Estimator Prob- ability Inputs Regressor Predict real no. Dec Tree, Sigmoid Perceptron, Sigmoid N.Net, Gauss/Joint BC, Gauss Naïve BC, N.Neigh, Bayes Net Based BC Joint DE, Naïve DE, Gauss/Joint DE, Gauss Naïve DE, Bayes Net Structure Learning Linear Regression, Quadratic Regression, Perceptron, Neural Net, N.Neigh, Kernel, LWR

Bayes Net Structure: Slide 12 Copyright © 2001, Andrew W. Moore And a new operation… Inputs Classifier Predict category Inputs Density Estimator Prob- ability Inputs Regressor Predict real no. Dec Tree, Sigmoid Perceptron, Sigmoid N.Net, Gauss/Joint BC, Gauss Naïve BC, N.Neigh, Bayes Net Based BC Joint DE, Naïve DE, Gauss/Joint DE, Gauss Naïve DE, Bayes Net Structure Learning Linear Regression, Quadratic Regression, Perceptron, Neural Net, N.Neigh, Kernel, LWR Inputs Inference Engine Learn P(E 1 |E 2 ) Joint DE, Bayes Net Structure Learning