Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming):

Slides:



Advertisements
Similar presentations
Mean-Field Theory and Its Applications In Computer Vision1 1.
Advertisements

Bayesian Belief Propagation
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Department of Computer Science Undergraduate Events More
Exact Inference in Bayes Nets
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
CSCI 121 Special Topics: Bayesian Networks Lecture #5: Dynamic Bayes Nets.
Loopy Belief Propagation a summary. What is inference? Given: –Observabled variables Y –Hidden variables X –Some model of P(X,Y) We want to make some.
An Introduction to Variational Methods for Graphical Models.
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
3 March, 2003University of Glasgow1 Statistical-Mechanical Approach to Probabilistic Inference --- Cluster Variation Method and Generalized Loopy Belief.
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
CS774. Markov Random Field : Theory and Application Lecture 04 Kyomin Jung KAIST Sep
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Distributed Message Passing for Large Scale Graphical Models Alexander Schwing Tamir Hazan Marc Pollefeys Raquel Urtasun CVPR2011.
Bayesian network inference
Belief Propagation in a Continuous World Andrew Frank 11/02/2009 Joint work with Alex Ihler and Padhraic Smyth TexPoint fonts used in EMF. Read the TexPoint.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Practical Belief Propagation in Wireless Sensor Networks Bracha Hod Based on a joint work with: Danny Dolev, Tal Anker and Danny Bickson The Hebrew University.
Global Approximate Inference Eran Segal Weizmann Institute.
CSI Uncertainty in A.I. Lecture Three Main Approaches To Approximate Inference MCMC Variational Methods Loopy belief propagation.
Conditional Random Fields
Belief Propagation, Junction Trees, and Factor Graphs
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
Belief Propagation Kai Ju Liu March 9, Statistical Problems Medicine Finance Internet Computer vision.
Computer vision: models, learning and inference
Active Learning for Networked Data Based on Non-progressive Diffusion Model Zhilin Yang, Jie Tang, Bin Xu, Chunxiao Xing Dept. of Computer Science and.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
6. Experimental Analysis Visible Boltzmann machine with higher-order potentials: Conditional random field (CRF): Exponential random graph model (ERGM):
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by B.-H. Kim Biointelligence Laboratory, Seoul National.
12/07/2008UAI 2008 Cumulative Distribution Networks and the Derivative-Sum-Product Algorithm Jim C. Huang and Brendan J. Frey Probabilistic and Statistical.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
GraphLab: how I understood it with sample code Aapo Kyrola, Carnegie Mellon Univ. Oct 1, 2009.
Belief Propagation. What is Belief Propagation (BP)? BP is a specific instance of a general class of methods that exist for approximate inference in Bayes.
Probabilistic Graphical Models
Physics Fluctuomatics / Applied Stochastic Process (Tohoku University) 1 Physical Fluctuomatics Applied Stochastic Process 9th Belief propagation Kazuyuki.
28 February, 2003University of Glasgow1 Cluster Variation Method and Probabilistic Image Processing -- Loopy Belief Propagation -- Kazuyuki Tanaka Graduate.
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
Solving Bayesian Decision Problems: Variable Elimination and Strong Junction Tree Methods Presented By: Jingsong Wang Scott Langevin May 8, 2009.
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 7th~10th Belief propagation Kazuyuki Tanaka Graduate School of Information Sciences,
Daphne Koller Message Passing Belief Propagation Algorithm Probabilistic Graphical Models Inference.
CS Statistical Machine learning Lecture 24
1 Mean Field and Variational Methods finishing off Graphical Models – Carlos Guestrin Carnegie Mellon University November 5 th, 2008 Readings: K&F:
Belief Propagation and its Generalizations Shane Oldenburger.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Distributed cooperation and coordination using the Max-Sum algorithm
10 October, 2007 University of Glasgow 1 EM Algorithm with Markov Chain Monte Carlo Method for Bayesian Image Analysis Kazuyuki Tanaka Graduate School.
Recap of L09: Normative Decision Theory
Nevin L. Zhang Room 3504, phone: ,
Maximum Expected Utility
StatSense In-Network Probabilistic Inference over Sensor Networks
Markov Networks.
Anytime Anyspace AND/OR Best-first Search for Bounding Marginal MAP
CSCI 5822 Probabilistic Models of Human and Machine Learning
Bucket Renormalization for Approximate Inference
Markov Random Fields Presented by: Vladan Radosavljevic.
≠ Particle-based Variational Inference for Continuous Systems
Physical Fluctuomatics 7th~10th Belief propagation
Expectation-Maximization & Belief Propagation
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
Readings: K&F: 11.3, 11.5 Yedidia et al. paper from the class website
Markov Networks.
Mean Field and Variational Methods Loopy Belief Propagation
Generalized Belief Propagation
Kazuyuki Tanaka Graduate School of Information Sciences
Presentation transcript:

Perfect recall: Every decision node observes all earlier decision nodes and their parents (along a “temporal” order) Sum-max-sum rule (dynamical programming): Perfect recall is unrealistic: memory limit, decentralized systems Variational methods: Log-partition function duality: Junction graph BP: approximating and Belief Propagation for Structured Decision Making Qiang Liu Alexander Ihler Department of Computer Science, University of California, Irvine Abstract Variational inference methods such as loopy BP have revolutionized inference abilities on graphical models. Influence diagrams (or decision networks) are extension of graphical models for representing structured decision making problems. Our contribution: A general variational framework for solving influence diagrams A junction graph belief propagation for IDs with an intuitive interpretation and strong theoretical guarantees A convergent double-loop algorithm Significant empirical improvement over the baseline algorithm Variational Framework for structured decision Influence Diagram Graphical Models and Variational Methods Graphical models: Factors & exponential family form Graphical representations: Bayes nets, Markov random fields … Inference: answering queries about graphical models Our Algorithms Experiments Junction graph belief propagation for MEU: Construct junction graph over the augmented distribution Main result: Intuition: the last term encourages policies to be deterministic Perfect recall  convex optimization (easier) Imperfect recall  non-convex optimization (harder) Bethe-Kikuchi approximation : locally consistent polytope ed abcbcd abe d bc e ab b e a c d Loopy Junction graph Influence diagram: Chance nodes ( C ): Augmented distribution: Maximum expected utility (MEU): Imperfect recall: No closed form solution Dominant algorithm: single policy updating (SPU), with policy-by- policy optimality If is the maximum, the optimal strategy is Causes policies to be deterministic Significance: Enables converting arbitrary variational methods to MEU algorithms “Integrates” the policy evaluation and policy improvement steps (avoiding expensive inner loops) c1c4d1c1c4d1 c1c2d2c1c2d2 c3d1c3d1 c2c3d3c2c3d3 c4d2d3c4d2d3 Influence diagram Augmented distribution (factor graph) Junction graph For each decision node, identify a unique cluster (called a decision cluster) that includes Decision cluster of d 1 Normal cluster Message passing algorithm ( ) Sum-messages (from normal clusters): MEU-messages (from decision clusters): Optimal policies: Strong local optimality: provably better than SPU Convergent algorithm by proximal point method: Iteratively optimize a smoothed objective, Diagnostic network (UAI08 inference challenge): e.g., calculating (log) partition function: Decentralized Sensor network: 1 Conditional probability: Decision rule: Global utility function: Local utility function: Decision nodes ( D ): Utility nodes ( U ): or d1d1 d2d2 u d1d1 d2d2 u Perfect recall Imperfect recall Additive d1d1 d2d2 utility Toy example: Multiplicative Weather Activity Forecast Happiness d3d3 d2d2 u c2c2 c3c3 d1d1 c4c4 c1c1 d3d3 d2d2 c2c2 c3c3 d1d1 c4c4 c1c1