Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.

Slides:



Advertisements
Similar presentations
Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Advertisements

Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Belief Networks Done by: Amna Omar Abdullah Fatima Mahmood Al-Hajjat Najwa Ahmad Al-Zubi.
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
Review: Bayesian learning and inference
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
PGM 2003/04 Tirgul 3-4 The Bayesian Network Representation.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Bayesian Network Representation Continued
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
. DAGs, I-Maps, Factorization, d-Separation, Minimal I-Maps, Bayesian Networks Slides by Nir Friedman.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
A Brief Introduction to Graphical Models
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Uncertainty Russell and Norvig: Chapter 14, 15 Koller article on BNs CMCS424 Spring 2002 April 23.
Artificial Intelligence CS 165A Thursday, November 29, 2007  Probabilistic Reasoning / Bayesian networks (Ch 14)
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
Aprendizagem Computacional Gladys Castillo, UA Bayesian Networks Classifiers Gladys Castillo University of Aveiro.
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
UIUC CS 598: Section EA Graphical Models Deepak Ramachandran Fall 2004 (Based on slides by Eyal Amir (which were based on slides by Lise Getoor and Alvaro.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Knowledge Representation & Reasoning Lecture #4 UIUC CS 498: Section EA Professor: Eyal Amir Fall Semester 2005 (Based on slides by Lise Getoor and Alvaro.
Conditional Probability Distributions Eran Segal Weizmann Institute.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
1 Bayesian Networks (Directed Acyclic Graphical Models) The situation of a bell that rings whenever the outcome of two coins are equal can not be well.
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Exploiting Structure in Probability Distributions Irit Gat-Viks Based on presentation and lecture notes of Nir Friedman, Hebrew University.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Reasoning Under Uncertainty: Independence and Inference CPSC 322 – Uncertainty 5 Textbook §6.3.1 (and for HMMs) March 25, 2011.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) Nov, 13, 2013.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Knowledge Representation & Reasoning Lecture #5 UIUC CS 498: Section EA Professor: Eyal Amir Fall Semester 2005 (Based on slides by Lise Getoor and Alvaro.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
Reasoning Under Uncertainty: Belief Networks
CS 2750: Machine Learning Directed Graphical Models
Bayesian networks Chapter 14 Section 1 – 2.
Qian Liu CSE spring University of Pennsylvania
Read R&N Ch Next lecture: Read R&N
Learning Bayesian Network Models from Data
Read R&N Ch Next lecture: Read R&N
Structure and Semantics of BN
Bayesian Networks Based on
Bayesian Networks Independencies Representation Probabilistic
Class #16 – Tuesday, October 26
Structure and Semantics of BN
Read R&N Ch Next lecture: Read R&N
Probabilistic Reasoning
Read R&N Ch Next lecture: Read R&N
Warm-up as you walk in Each node in a Bayes net represents a conditional probability distribution. What distribution do you get when you multiply all of.
Probabilistic Reasoning
Presentation transcript:

Bayesian Networks VISA Hyoungjune Yi

BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert System

Common Sense Reasoning about uncertainty June is waiting for Larry and Jacobs who are both late for VISA seminar June is worried that if the roads are icy one or both of them may have crash his car Suddenly June learns that Larry has crashed June think: “If Larry has crashed then probably the roads are icy. So Jacobs has also crashed” June then learns that it is warm outside and roads are salted June Think: “Larry was unlucky; Jacobs should still make it”

Causal Relationships State of Road Icy/ not icy Jacobs Crash/No crash Larry Crash/No crash

Larry Crashed ! State of Road Icy/ not icy Jacobs Crash/No crash Larry Crash/No crash Information Flow

But Roads are dry State of Road not icy Jacobs Crash/No crash Larry Crash/No crash Information Flow

Wet grass To avoid icy roads, Larry moves to UCLA; Jacobs moves in USC One morning as Larry leaves for work, he notices that his grass is wet. He wondered whether he has left his sprinkler on or it has rained Glancing over to Jacobs’ lawn he notices that it is also get wet Larry thinks: “Since Jacobs’ lawn is wet, it probably rained last night” Larry then thinks: “If it rained then that explains why my lawn is wet, so probably the sprinkler is off”

Larry’s grass is wet Rain Yes/no Larry’s grass Wet Jacobs grass Wet/Dry Information Flow Sprinkler On/Off

Jacobs’ grass is also wet Rain Yes/no Larry’s grass Wet Jacobs grass Wet Information Flow Sprinkler On/Off

Bayesian Network Data structure which represents the dependence between variables Gives concise specification of joint prob. dist. Bayesian Belief Network is a graph that holds – Nodes are a set of random variables – Each node has a conditional prob. Table – Edges denote conditional dependencies – DAG : No directed cycle – Markov condition

Bayesian network Markov Assumption – Each random variable X is independent of its non- descendent given its parent Pa(X) – Formally, Ind(X; NonDesc(X) | Pa(X)) if G is an I-MAP of P (<-? ) I-MAP? Later X Y1Y1 Y2Y2

Markov Assumption In this example: – Ind( E; B ) – Ind( B; E, R ) – Ind( R; A, B, C | E ) – Ind( A; R | B, E ) – Ind( C; B, E, R | A) Earthquake Radio Burglary Alarm Call

I-Maps A DAG G is an I-Map of a distribution P if the all Markov assumptions implied by G are satisfied by P Examples: XYXY

I-MAP G is Minimal I-Map iff – G is I-Map of P – If G’  G then G’ is not an I-Map of P I-Map is not unique

Factorization Given that G is an I-Map of P, can we simplify the representation of P ? Example: Since Ind(X;Y), we have that P(X|Y) = P(X) Applying the chain rule P(X,Y) = P(X|Y) P(Y) = P(X) P(Y) Thus, we have a simpler representation of P(X,Y) XY

Factorization Theorem Thm: if G is an I-Map of P, then P(C,A,R,E,B) = P(B)P(E|B)P(R|E,B)P(A|R,B,E)P(C|A,R,B,E) versus P(C,A,R,E,B) = P(B) P(E) P(R|E) P(A|B,E) P(C|A) Earthquake Radio Burglary Alarm Call

So, what ? We can write P in terms of “local” conditional probabilities If G is sparse, that is, |Pa(Xi)| < k,  each conditional probability can be specified compactly e.g. for binary variables, these require O(2k) params.  representation of P is compact linear in number of variables

Formal definition of BN A Bayesian network specifies a probability distribution via two components: – A DAG G – A collection of conditional probability distributions P(Xi|Pai) The joint distribution P is defined by the factorization Additional requirement: G is a minimal I-Map of P

Bayesian Network - Example Each node X i has a conditional probability distribution P(X i |Pa i ) – If variables are discrete, P is usually multinomial – P can be linear Gaussian, mixture of Gaussians, … XRay Lung Infiltrates Sputum Smear TuberculosisPneumonia p t p tp t t p T P P(I |P, T )

BN Semantics Compact & natural representation: – nodes have  k parents  2k n vs. 2n params conditional independencies in BN structure + local probability models full joint distribution over domain = X I S TP

d-separation d-sep(X;Y | Z, G) – X is d-separated from Y, given Z if all paths from a node in X to a node in Y are blocked given Z Meaning ? – On the blackboard – Path Active: dependency between end nodes in the path Blocked: No dependency – Common cause, Intermediate, common effect On the blackboard

BN – Belief, Evidence and Query BN is for “Query” - partly Query involves evidence – Evidence is an assignment of values to a set of variables in the domain Query is a posteriori belief – Belief P(x) = 1 or P(x) = 0

Learning Structure Problem Definition – Given: Data D – Return: directed graph expressing BN Issue – Superfluous edges – Missing edges Very difficult –

BN Learning BN models can be learned from empirical data – parameter estimation via numerical optimization – structure learning via combinatorial search. BN hypothesis space biased towards distributions with independence structure. Induce r Data X I S TP