Solving Bayesian Decision Problems: Variable Elimination and Strong Junction Tree Methods Presented By: Jingsong Wang Scott Langevin May 8, 2009.

Slides:



Advertisements
Similar presentations
Heuristic Search techniques
Advertisements

Tree Clustering for Constraint Networks 1 Chris Reeson Advanced Constraint Processing Fall 2009 By Rina Dechter & Judea Pearl Artificial Intelligence,
Minimum Clique Partition Problem with Constrained Weight for Interval Graphs Jianping Li Department of Mathematics Yunnan University Jointed by M.X. Chen.
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
CS498-EA Reasoning in AI Lecture #15 Instructor: Eyal Amir Fall Semester 2011.
Graphical Models - Inference - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
1 Chapter 5 Belief Updating in Bayesian Networks Bayesian Networks and Decision Graphs Finn V. Jensen Qunyuan Zhang Division. of Statistical Genomics,
Bayesian Networks, Winter Yoav Haimovitch & Ariel Raviv 1.
Lauritzen-Spiegelhalter Algorithm
Copyright 2004 Koren & Krishna ECE655/DataRepl.1 Fall 2006 UNIVERSITY OF MASSACHUSETTS Dept. of Electrical & Computer Engineering Fault Tolerant Computing.
Statistical Methods in AI/ML Bucket elimination Vibhav Gogate.
Exact Inference in Bayes Nets
THE HONG KONG UNIVERSITY OF SCIENCE & TECHNOLOGY CSIT 5220: Reasoning and Decision under Uncertainty L09: Graphical Models for Decision Problems Nevin.
MINIMUM COST FLOWS: NETWORK SIMPLEX ALGORITHM A talk by: Lior Teller 1.
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering A Comparison of Lauritzen- Spiegelhalter, Hugin, and Shenoy- Shafer Architectures.
Overview of Inference Algorithms for Bayesian Networks Wei Sun, PhD Assistant Research Professor SEOR Dept. & C4I Center George Mason University, 2009.
Hidden Markov Models M. Vijay Venkatesh. Outline Introduction Graphical Model Parameterization Inference Summary.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
From Variable Elimination to Junction Trees
CSCI 121 Special Topics: Bayesian Networks Lecture #3: Multiply-Connected Graphs and the Junction Tree Algorithm.
Introduction to Inference for Bayesian Netoworks Robert Cowell.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
CS774. Markov Random Field : Theory and Application Lecture 06 Kyomin Jung KAIST Sep
On the Role of MSBN to Cooperative Multiagent Systems By Y. Xiang and V. Lesser Presented by: Jingshan Huang and Sharon Xi.
A Graphical Model For Simultaneous Partitioning And Labeling Philip Cowans & Martin Szummer AISTATS, Jan 2005 Cambridge.
Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011.
Branch and Bound Similar to backtracking in generating a search tree and looking for one or more solutions Different in that the “objective” is constrained.
April 2004Bayesian Networks in Educational Assessment - Section II Tutorial 12 Inference Based on ETS lecture.
Bayesian Networks Clique tree algorithm Presented by Sergey Vichik.
Belief Propagation, Junction Trees, and Factor Graphs
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Learning Equivalence Classes of Bayesian-Network Structures David M. Chickering Presented by Dmitry Zinenko.
Mobile and Wireless Computing Institute for Computer Science, University of Freiburg Western Australian Interactive Virtual Environments Centre (IVEC)
Exact Inference: Clique Trees
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Factor Graphs Young Ki Baik Computer Vision Lab. Seoul National University.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Belief Propagation. What is Belief Propagation (BP)? BP is a specific instance of a general class of methods that exist for approximate inference in Bayes.
Introduction to Bayesian Networks
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 11 th, 2006 Readings: K&F: 8.1, 8.2, 8.3,
Daphne Koller Message Passing Belief Propagation Algorithm Probabilistic Graphical Models Inference.
An Introduction to Variational Methods for Graphical Models
Intro to Junction Tree propagation and adaptations for a Distributed Environment Thor Whalen Metron, Inc.
Probabilistic Graphical Models seminar 15/16 ( ) Haim Kaplan Tel Aviv University.
Problem Reduction So far we have considered search strategies for OR graph. In OR graph, several arcs indicate a variety of ways in which the original.
Indexing Correlated Probabilistic Databases Bhargav Kanagal, Amol Deshpande University of Maryland, College Park, USA SIGMOD Presented.
1 Use graphs and not pure logic Variables represented by nodes and dependencies by edges. Common in our language: “threads of thoughts”, “lines of reasoning”,
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Arc Consistency CPSC 322 – CSP 3 Textbook § 4.5 February 2, 2011.
1 An Arc-Path Model for OSPF Weight Setting Problem Dr.Jeffery Kennington Anusha Madhavan.
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Graph Connectivity This discussion concerns connected components of a graph. Previously, we discussed depth-first search (DFS) as a means of determining.
Today Graphical Models Representing conditional dependence graphically
Belief propagation with junction trees Presented by Mark Silberstein and Yaniv Hamo.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 15 th, 2008 Readings: K&F: 8.1, 8.2, 8.3,
Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming):
5.6 Prefix codes and optimal tree Definition 31: Codes with this property which the bit string for a letter never occurs as the first part of the bit string.
Alternative Search Formulations and Applications
1 R-Trees Guttman. 2 Introduction Range queries in multiple dimensions: Computer Aided Design (CAD) Geo-data applications Support special data objects.
Recap of L09: Normative Decision Theory
Nevin L. Zhang Room 3504, phone: ,
Constraint Propagation
CSCI 5822 Probabilistic Models of Human and Machine Learning
Graphs Chapter 11 Objectives Upon completion you will be able to:
Lecture 3: Exact Inference in GMs
Compact routing schemes with improved stretch
Presentation transcript:

Solving Bayesian Decision Problems: Variable Elimination and Strong Junction Tree Methods Presented By: Jingsong Wang Scott Langevin May 8, 2009

Introduction Solutions to Influence Diagrams Variable Elimination Strong Junction Tree Hugin Architecture Conclusions

3 Solutions to Influence Diagrams

4 The example influence diagram, DI I 0 = Φ, I 1 = {T}, I 2 = {A, B, C} Chance Nodes Decision Nodes Utility Nodes

5 Solutions to Influence Diagrams

6

7 The Chain Rule for Influence Diagrams

8 Strategies and Expected Utilities

9 DI unfolded into a decision tree. Apply average-out and fold-back algorithm. To reduce the size of the decision tree the last chance node in each path is defined as the Cartesian product of A and C, and that the utilities in the leaves are the sums of V 1 and V 2. Strategies and Expected Utilities

10 Strategies and Expected Utilities – D2

11 Strategies and Expected Utilities – D2

12 Strategies and Expected Utilities – D2 The decision tree with D 2 replaced by a utility function reflecting that the policy δ 2 for D 2 is followed.

13 Strategies and Expected Utilities – D1

14 Strategies and Expected Utilities - Combined D1

15 Strategies and Expected Utilities - Combined D2

16 Strategies and Expected Utilities

17 Strategies and Expected Utilities - Proof

18 Strategies and Expected Utilities - Proof

19 Strategies and Expected Utilities - Proof

20 Variable Elimination Compare the method for solving influence diagrams with the junction tree propagation algorithm –Similarities: Start off with a set of potentials Eliminate one variable at a time –Differences: The elimination order is constrained by the temporal order Two types of potentials to deal with Need to eliminate in only one direction Strong elimination order –Sum-marginalize I n, then max-marginalize D n, sum-marginalize I n-1, etc

21 Variable Elimination Analyze the calculations in eliminating a variable –Φ a set of probability potentials –Ψ a set of utility potentials –The product of all probability potentials multiplied by the sum of all utility potentials:

22 Variable Elimination – Sum-Marginalization

23 Variable Elimination – Max-Marginalization

24 Variable Elimination

Strong Junction Tree Methods Rely on secondary computational structure to calculate MEU and policies Similar idea of Junction Tree Method, only here the order of elimination is constrained by the partial order Hugin Method Lazy Propagation Method Creating a Strong Junction Tree –Moralize Influence Graph –Triangulate Moralized Graph –Arrange Cliques into Strong Junction Tree

Running Example Partial Temporal Order: I 0 = {B}, D 1, I 1 = {E,F}, D 2, I 2 = Ø, D 3, I 3 = {G}, D 4, I 4 = {A, C, D, H, I, J, K, L}

Moralization of Influence Diagram Remove informational links Add a link between nodes with a common child Remove utility nodes Remove directional arrows

Moralization of Influence Diagram →

Strong Triangulation of Moral Graph Triangulate by eliminating nodes from moral graph according to reverse of partial order imposed by influence diagram: Nodes in I k have no imposed order and can be eliminated in any order (ex: use min fill-in heuristic)

Strong Triangulation of Moral Graph → Partial Order: I 0 = {B}, D 1, I 1 = {E,F}, D 2, I 2 = Ø, D 3, I 3 = {G}, D 4, I 4 = {A, C, D, H, I, J, K, L} Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Strong Junction Tree Construction Organize cliques of triangulated graph into a strong junction tree: –Each pair of cliques (C1, C2), C1 ∩ C2 is contained in every clique on the path connecting C1 and C2 –Each pair of cliques (C1, C2) with C1 closer to root R than C2, there exists an ordering of variables in C2 that respects the partial order and with the nodes in C1 ∩ C2 preceding the variables in C2\C1 Ensures that the maximum expected utility can be computed via local message passing in the junction tree

Strong Junction Tree Construction Algorithm for generation of Strong Junction Tree Number nodes in triangulated graph according to reverse of elimination order chosen during triangulation Let C be a clique in the triangulated graph and v be the highest numbered node in C with neighbor node u not in C and the number associated with u < the number associated with v If such a node v exists, then set index of C to number of v, else set index of C to 1 Order the cliques in increasing order according to their index This order will have the running intersection property: To construct Strong Junction Tree: start with C 1 (the root) then successively attach each clique C k to some clique C j that contains S k

Strong Junction Tree Construction → Partial Order: I 0 = {B}, D 1, I 1 = {E,F}, D 2, I 2 = Ø, D 3, I 3 = {G}, D 4, I 4 = {A, C, D, H, I, J, K, L} Cliques: {B,D 1,E,F,D}, {B,C,A}, {B,E,D,C}, {E,D 2,G}, {D 2,G,D 4,I}, {D 4,I,L}, {D 3,H,K}, {H,K,J}, {F,D 3,H}

Hugin Architecture Each clique C and separator S in Junction Tree contains a probability potential ϕ and a utility potential Ψ Initialize Junction Tree –Assign each potential ϕ to one and only one clique C where dom( ϕ ) ⊆ dom(C) –Combine potentials in each clique: –Assign unity potential to cliques with no probability potential assigned –Assign null utility potential to cliques with no utility potential assigned

Hugin Architecture Uses message passing in the strong junction tree Messages are passed from leave nodes towards the root node via adjacent nodes A clique node can pass a message when it has received messages from all adjacent nodes further from the root node Messages are stored in the separator S connecting two adjacent nodes The message consists of two potentials: probability potential ϕ S and a utility potential Ψ S that are calculated as: Note that ∑ is a general marginalization operator and is a summation for probability nodes and a max function for decision nodes. Nodes are marginalized according to reverse of partial order Message from C j is absorbed by C i by:

Hugin Architecture The optimal policy for a decision variable can be determined from the potentials of the clique or separator that is closest to root and contains the decision variable (it may be the root itself) The MEU is calculated using the potentials in the root node after message passing has completed

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 {U 4 (L)}, {P(L|D 4,I)} {P(I|D 2,G)} {P(G|E)} {U 1 (D 1 )}, {P(F|D), P(D|B,D 1 ), P(B)} {P(E|C,D)} {P(C|A,B), P(A)} {U 3 (J, K)}, {P(J|H)} {P(K|D 3,H)} {U 2 (D 3 )}, {P(H|F)} Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Ψ(L), ϕ (L,D 4,I) ϕ (I,D 2,G) ϕ (G,E) Ψ(D 1 ), ϕ (B,D,F, D 1 ) ϕ (E,C,D) ϕ (C,A,B) Ψ(J,K), ϕ (J,H) ϕ (K,D 3,H) Ψ(D 3 ), ϕ (H,F) Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B

Hugin Architecture C1 C2C3C4 C5C6C7 C8C9 Elimination Order: A, L, I, J, K, H, C, D, D 4, G, D 3, D 2, E, F, D 1, B Calculate Policy: D 1 use C1 D 2 use C2 D 3 use C3 D 4 use C5 MEU: Use C1

Conclusion We reviewed two methods for solving influence diagrams: –Variable Elimination –Strong Junction Tree Method (Hugin) There are other methods that were not discussed: –Lazy Propagation –Node Removal and Arc Reversal

Questions?