Local structures; Causal Independence, Context-sepcific independance COMPSCI 276 Fall 2007.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Reasoning Under Uncertainty: Bayesian networks intro Jim Little Uncertainty 4 November 7, 2014 Textbook §6.3, 6.3.1, 6.5, 6.5.1,
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.
Probabilistic networks Inference and Other Problems Hans L. Bodlaender Utrecht University.
Introduction of Probabilistic Reasoning and Bayesian Networks
Pearl’s Belief Propagation Algorithm Exact answers from tree-structured Bayesian networks Heavily based on slides by: Tomas Singliar,
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Bayesian networks Chapter 14 Section 1 – 2.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
. Inference I Introduction, Hardness, and Variable Elimination Slides by Nir Friedman.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Exact Inference Eran Segal Weizmann Institute. Course Outline WeekTopicReading 1Introduction, Bayesian network representation1-3 2Bayesian network representation.
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graphs.
. DAGs, I-Maps, Factorization, d-Separation, Minimal I-Maps, Bayesian Networks Slides by Nir Friedman.
Probability and Statistics Review Thursday Sep 11.
Inference in Gaussian and Hybrid Bayesian Networks ICS 275B.
CS B 553: A LGORITHMS FOR O PTIMIZATION AND L EARNING Continuous Probability Distributions and Bayesian Networks with Continuous Variables.
Probabilistic Reasoning
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
Bayesian Belief Network Compiled By: Raj Gaurang Tiwari Assistant Professor SRMGPC, Lucknow.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Reasoning Under Uncertainty: Bayesian networks intro CPSC 322 – Uncertainty 4 Textbook §6.3 – March 23, 2011.
Artificial Intelligence CS 165A Thursday, November 29, 2007  Probabilistic Reasoning / Bayesian networks (Ch 14)
1 CS 391L: Machine Learning: Bayesian Learning: Beyond Naïve Bayes Raymond J. Mooney University of Texas at Austin.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Baye’s Rule.
Automated Planning and Decision Making Prof. Ronen Brafman Automated Planning and Decision Making 2007 Bayesian networks Variable Elimination Based on.
Direct Message Passing for Hybrid Bayesian Networks Wei Sun, PhD Assistant Research Professor SFL, C4I Center, SEOR Dept. George Mason University, 2009.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Lectures 2 – Oct 3, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall.
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 12th Bayesian network and belief propagation in statistical inference Kazuyuki Tanaka.
CS 416 Artificial Intelligence Lecture 14 Uncertainty Chapters 13 and 14 Lecture 14 Uncertainty Chapters 13 and 14.
Conditional Probability Distributions Eran Segal Weizmann Institute.
Math 4030 – 6a Joint Distributions (Discrete)
Lecture 2: Statistical learning primer for biologists
Wei Sun and KC Chang George Mason University March 2008 Convergence Study of Message Passing In Arbitrary Continuous Bayesian.
Lecture 29 Conditional Independence, Bayesian networks intro Ch 6.3, 6.3.1, 6.5, 6.5.1,
Representation of CPTs CH discrete Canonical distribution: standard Deterministic nodes: values computable exactly from parent nodes Noisy-OR relations:
1 Param. Learning (MLE) Structure Learning The Good Graphical Models – Carlos Guestrin Carnegie Mellon University October 1 st, 2008 Readings: K&F:
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
CS 416 Artificial Intelligence Lecture 15 Uncertainty Chapters 13 and 14 Lecture 15 Uncertainty Chapters 13 and 14.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Reasoning Patterns Bayesian Networks Representation Probabilistic
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
Daphne Koller Independencies Bayesian Networks Probabilistic Graphical Models Representation.
Probability and Information Theory
Context-Specific CPDs
I-equivalence Bayesian Networks Representation Probabilistic Graphical
Probabilistic Influence & d-separation
Reasoning Patterns Bayesian Networks Representation Probabilistic
Tree-structured CPDs Local Structure Representation Probabilistic
Representation Probabilistic Graphical Models Local Structure Overview.
Chapter 14 February 26, 2004.
Presentation transcript:

Local structures; Causal Independence, Context-sepcific independance COMPSCI 276 Fall 2007

Local structure2 Reducing parameters of families Determinizm Causal independence Context-specific independanc Continunous variables

Local structure3

4 Causal Independence Event X has two possible causes: A,B. It is hard to elicit P(X|A,B) but it is easy to determine P(X|A) and P(X|B). Example: several diseases causes a symptom. Effect of A on X is independent from the effect of B on X  Causal Independence, using canonical models: Noisy-O, Noisy AND, noisy-max AB X

Local structure5 Binary OR AB X ABP(X=0|A,B) 001 P(X=1|A,B)

Local structure6 Noisy-OR “noise” is associated with each edge described by noise parameter  [0,1] : Let q b =0.2, q a =0.1 P(x=0|a,b)= (1- a ) (1- b ) P(x=1|a,b)=1-(1- a ) (1- b ) AB X ABP(X=0|A,B) 001 P(X=1|A,B) 0 a b qi=P(X=0|A_i=1,…else =0)

Local structure7 Noisy-OR with Leak Use leak probability 0  [0,1] when both parents are false: Let a =0.2, b =0.1, 0 = P(x=0|a,b)= (1- 0 )(1- a ) a (1- b ) b P(x=0|a,b)=1-(1- 0 )(1- a ) a (1- b ) b AB X ABP(X=0|A,B) P(X=1|A,B) a b

Local structure8 Formal Definition for Noisy-Or Definition 1 Let Y be a binary-valued random variable with k binary-valued parents X 1,…,X k. The CPT P(Y|X 1,…X k ) is a noisy-or if there are k+1 noise parameters 0, 1,… k such that P(y=0| X 1,…,X k ) = (1- 0 )  i,Xi=1 (1- i )

Local structure9 Closed Form Bel(X) - 1 Given: noisy-or CPT P(x|u) noise parameters i T u = {i: U i = 1} Define: q i = 1 - I, Then: q_i is the probability that the inhibitor for u_i is active while the

Local structure10 Closed Form Bel(X) - 2 Using Iterative Belief Propagation : Set pi ix = pi x (u k =1). Then we can show that:

Local structure11 Closed Form Bel(X) - 2 Using Iterative Belief Propagation : Set pi ix = pi x (u k =1). Then we can show that:

Local structure12 Causal Influence Defined Definition 2 Let Y be a random variable with k parents X 1,…,X k. The CPT P(Y|X 1,…X k ) exhibits independence of causal influence (ICI) if it is described via a network fragment of the structure shown in on the left where CPT of Z is a deterministic functions f. Z Y X1X1 X1X1 X1X1 Z0Z0 Z1Z1 Z2Z2 ZkZk

Local structure13

Local structure14

Local structure15

Local structure16

Local structure17

Local structure18 Context Specific Independence When there is conditional independence in some specific variable assignment

Local structure19

Local structure20

Local structure21

Local structure22

Local structure23 The impact during inference Causal independence in polytrees is linear during inference Causal independence in general can sometime be exploited but not always CSI can be exploited by using operation (product and summation) over trees.

Local structure24 Representing CSI Using decision trees Using decision graphs

Local structure25 IntelligenceDifficulty Grade Letter SAT Job Apply A student’s example

Local structure26 A S L (0.8,0.2) (0.9,0.1)(0.4,0.6) (0.1,0.9) s1s1 a0a0 a1a1 s0s0 l1l1 l0l0 Tree CPD If the student does not apply, SAT and L are irrelevant Tree-CPD for job

Local structure27 Definition of CPD-tree A CPD-tree of a CPD P(Z|pa_Z) is a tree whose leaves are labeled by P(Z) and internal nodes correspond to parents branching over their values.

Local structure28 C L2 (0.1,0.9) l2 1 c1c1 c2c2 l2 0 L1 (0.8,0.2)(0.3,0.7) l1 1 l1 0 (0.9,0.1) Letter1 Job Letter2 Choice Captures irrelevant variables

Local structure29 Multiplexer CPD A CPD P(Y|A,Z1,Z2,…,Zk) is a multiplexer iff Val(A)=1,2,…k, and P(Y|A,Z1,…Zk)=Z_a Letter1 Letter Letter2 Choice Job

Local structure30 A B C (0.3,0.7)(0.4,0.6) (0.1,0.9) b1b1 a0a0 a1a1 b0b0 c1c1 c0c0 C B (0.3,0.7)(0.5,0.5) (0.2,0.8) c1c1 c0c0 b1b1 b0b0 Rule-based representation A CPD-tree that correponds to rules.

Continuous Variables ICS 275b 2002

Local structure32 Gaussian Distribution N( ,  )

Local structure gaussian(x,0,1) gaussian(x,1,1) N( ,  )

Local structure gaussian(x,0,1) gaussian(x,0,2) N( ,  )

Local structure35 Multivariate Gaussian Definition: Let X 1,…,X n. Be a set of random variables. A multivariate Gaussian distribution over X 1,…,X n is a parameterized by an n-dimensional mean vector  and an n x n positive definitive covariance matrix . It defines a joint density via:

Local structure36 Linear Gaussian Distribution Definition: Let Y be a continuous node with continuous parents X 1,…,X k. We say that Y has a linear Gaussian model if it can be described using parameters  0, …,  k and  2 such that: P(y| x 1,…,x k )=N (  0 +  1 x 1 +…,  k x k ;  2 )

Local structure37 XY XYXY

Local structure38

Local structure39 Linear Gaussian Network Definition Linear Gaussian Bayesian network is a Bayesian network all of whose variables are continuous and where all of the CPTs are linear Gaussians. Linear Gaussian BN  Multivariate Gaussian =>Linear Gaussian BN has a compact representation

Local structure40 Hybrid Models Continuous Node, Discrete Parents (CLG) –Define density function for each instantiation of parents Discrete Node, Continuous Parents –Treshold –Sigmoid

Local structure41 Continuous Node, Discrete Parents Definition: Let X be a continuous node, and let U={U 1,U 2,…,U n } be its discrete parents and Y={Y 1,Y 2,…,Y k } be its continuous parents. We say that X has a conditional linear Gaussian (CLG) CPT if, for every value u  D(U), we have a a set of (k+1) coefficients a u,0, a u,1, …, a u,k+1 and a variance  u 2 such that:

Local structure42 CLG Network Definition: A Bayesian network is called a CLG network if every discrete node has only discrete parents, and every continuous node has a CLG CPT.

Local structure43 Discrete Node, Continuous Parents Threshold Model

Local structure44 Discrete Node, Continuous Parents Sigmoid Binomial Logit Definition: Let Y be a binary-valued random variable with k continuous-valued parents X 1,…X k. The CPT P(Y|X 1 …X k ) is a linear sigmoid (also called binomial logit) if there are (k+1) weights w 0,w 1,…,w k such that:

Local structure sigmoid(0.1*x) sigmoid(0.5*x) sigmoid(0.9*x)

Local structure46 References Judea Pearl “Probabilistic Reasoning in Inteeligent Systems”, section 4.3 Nir Friedman, Daphne Koller “Bayesian Network and Beyond”