Bayes Nets. Bayes Nets Quick Intro Topic of much current research Models dependence/independence in probability distributions Graph based - aka “graphical.

Slides:



Advertisements
Similar presentations
Introduction to Graphical Models Brookes Vision Lab Reading Group.
Advertisements

Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for 1 Lecture Notes for E Alpaydın 2010.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Lauritzen-Spiegelhalter Algorithm
Lirong Xia Bayesian networks (2) Thursday, Feb 25, 2014.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
Lirong Xia Hidden Markov Models Tue, March 28, 2014.
Homework 3: Naive Bayes Classification
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
Bayes for beginners Methods for dummies 27 February 2013 Claire Berna
Introduction of Probabilistic Reasoning and Bayesian Networks
Pearl’s Belief Propagation Algorithm Exact answers from tree-structured Bayesian networks Heavily based on slides by: Tomas Singliar,
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Bayesian Belief Networks
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Bayesian Networks Alan Ritter.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Bayesian Networks Textbook: Probabilistic Reasoning, Sections 1-2, pp
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Bayesian networks Chapter 14. Outline Syntax Semantics.
A Brief Introduction to Graphical Models
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Perceptual and Sensory Augmented Computing Machine Learning, Summer’11 Machine Learning – Lecture 13 Introduction to Graphical Models Bastian.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Probabilistic Reasoning ECE457 Applied Artificial Intelligence Spring 2007 Lecture #9.
Supervised Learning Approaches Bayesian Learning Neural Network Support Vector Machine Ensemble Methods Adapted from Lecture Notes of V. Kumar and E. Alpaydin.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Bayesian Networks Aldi Kraja Division of Statistical Genomics.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
Bayesian Decision Theory Introduction to Machine Learning (Chap 3), E. Alpaydin.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
Lecture 2: Statistical learning primer for biologists
CIAR Summer School Tutorial Lecture 1b Sigmoid Belief Nets Geoffrey Hinton.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Quiz 3: Mean: 9.2 Median: 9.75 Go over problem 1.
Pattern Recognition and Machine Learning
Bayes network inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y 
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Bayesian Decision Theory Introduction to Machine Learning (Chap 3), E. Alpaydin.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
Integrative Genomics I BME 230. Probabilistic Networks Incorporate uncertainty explicitly Capture sparseness of wiring Incorporate multiple kinds of data.
INTRODUCTION TO Machine Learning 2nd Edition
Read R&N Ch Next lecture: Read R&N
INTRODUCTION TO Machine Learning
Markov Networks.
Read R&N Ch Next lecture: Read R&N
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CAP 5636 – Advanced Artificial Intelligence
Markov Networks.
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
Approximate Inference by Sampling
Lecture 3: Exact Inference in GMs
INTRODUCTION TO Machine Learning
Read R&N Ch Next lecture: Read R&N
Bayesian networks (2) Lirong Xia. Bayesian networks (2) Lirong Xia.
Bayesian networks (2) Lirong Xia.
Presentation transcript:

Bayes Nets

Bayes Nets Quick Intro Topic of much current research Models dependence/independence in probability distributions Graph based - aka “graphical models” “belief nets” Two kinds - directed and undirected

Basic idea: Model a joint distribution as a DAG with attributes (features) at the nodes Sample distribution from sources Party Apti- tude study Pass Exam Sleep

At each node is a conditional distribution for the attribute depending only on incoming arcs Party study P(yes) =.6 P(no) =.4 Party yes P(yes) = 0.1 P(no)=0.9 Party No P(yes)=0.9 P(no) = 0.1)

Causes and Bayes’ Rule Diagnostic inference: Knowing that the grass is wet, what is the probability that rain is the cause? causal diagnostic

Causal vs Diagnostic Inference Causal inference: If the sprinkler is on, what is the probability that the grass is wet? P(W|S) = P(W|R,S) P(R|S) + P(W|~R,S) P(~R|S) = P(W|R,S) P(R) + P(W|~R,S) P(~R) = = 0.92 Diagnostic inference: If the grass is wet, what is the probability that the sprinkler is on? P(S|W) = 0.35 > 0.2 P(S) P(S|R,W) = 0.21Explaining away: Knowing that it has rained decreases the probability that the sprinkler is on.

Bayesian Networks: Causes Causal inference: P(W|C) = P(W|R,S) P(R,S|C) + P(W|~R,S) P(~R,S|C) + P(W|R,~S) P(R,~S|C) + P(W|~R,~S) P(~R,~S|C) and use the fact that P(R,S|C) = P(R|C) P(S|C) Diagnostic: P(C|W ) = ?

Bayesian Nets: Local structure P (F | C) = ?

Bayesian Networks: Inference P (C,S,R,W,F) = P (C) P (S|C) P (R|C) P (W|R,S) P (F|R) P (C,F) = ∑ S ∑ R ∑ W P (C,S,R,W,F) P (F|C) = P (C,F) / P(C)Not efficient! Belief propagation (Pearl, 1988) Junction trees (Lauritzen and Spiegelhalter, 1988)

Conditional independence Know values of some nodes, which others are independent? –Nodes depend on ancestors through parents –Node conditionally indep of rest of graph if its parents, children, and children’s parents known –Notion of d-path - two nodes dependent if there is a d-path (or coincidence), independent otherwise

D-paths go through Linear non-evidence nodes (both ways) Diverging non-evid. Nodes converging evidence nodes: Conveging non-evidence w. evidence descd.

In general, working with Bayes nets hard, (easy for trees, also approximate inference - loopy propagation, sampling) Concise representation of some distributions Very flexible - can fix or learn structure, can have unobserved nodes (latent variables) Can capture conditional dependencies Generalizes HMMs and other models

Bayes nets problems: Full Observability Partial Observability Graph Known Maximum likelihood Expectation- Maximization Graph Unknown Local search over models Hard

Bayes Nets Applications Add probabilities to expert systems Vista system (Horvitz) for space shuttle problem detection/remedies Used in Microsoft products: Office ‘95 Answer Wizard, Office ‘97 Office Assistant, etc. Computational Biology Special cases applied to speech recognition, turbo coding, and many other places.