Download presentation
Presentation is loading. Please wait.
Published byTeresa Poole Modified over 9 years ago
1
. Expressive Graphical Models in Variational Approximations: Chain-Graphs and Hidden Variables Tal El-Hay & Nir Friedman School of Computer Science & Engineering Hebrew University
2
Inference in Graphical Models Exact Inference NP-hard, in general Can be efficient for certain classes What do we do when exact inference is intractable? Resort to approximate methods Approximate inference is also NP-hard But, specific approximation methods work for specific classes of models Need to enrich approximate methods
3
Variational Approximations u Approximate the posterior of a complex model using a simpler distribution u Choice of a simpler model method : Mean field, Structured approximations, and Mixture models
4
Variational Approximations u Approximate the posterior of a complex model using a simpler distribution u Choice of a simpler model method : Mean field, Structured approximations, and Mixture models
5
Variational Approximations u Approximate the posterior of a complex model using a simpler distribution u Choice of a simpler model method : Mean field, Structured approximations, and Mixture models
6
Variational Approximations u Approximate the posterior of a complex model using a simpler distribution u Choice of a simpler model method : Mean field, Structured approximations, and Mixture models
7
Enhancing Variational Approximations Basic tradeoff: accuracy complexity Goal: New families of approximating distributions better tradeoff
8
Outline u Structured variational approximations [review] u Using chain-graphs u Adding hidden variables u Discussion
9
Structured Approximations Target model: Approximation: where
10
Structured Approximations Goal: Maximize the following functional F[Q] is a lower bound on the log likelihood If Q is tractable then F[Q] might be tractable KL Distance 0
11
Structured Approximations u To characterize the maximum point we define the generalized functional u Differentiation yields the following equation approximates using the lower bound on the local distribution
12
Structured Approximations Optimization u Asynchronous updates guaranties convergence u Efficient calculation of the update formulas:
13
Chain Graph Approximations u Posterior distributions can be modeled as chain graphs = where
14
Chain Graph Approximations Chain graph distributions: where are potential functions on subsets of T u Generalize both Bayesian networks and Markov networks u A simple approximation example:
15
Chain Graph Approximations Optimization where
16
Adding Hidden Variables Potential pitfall: Multi-modal distributions u Jaakkola & Jordan: Use mixture models Modeling assumption: Factorized mixture components Generalization: Structured approximation with an extra set of hidden variables u Approximating distribution:
17
Adding Hidden Variables: Intuition Lower bound improvement potential where I(T;V) is the mutual information u Capture correlations in a compact manner:
18
Adding Hidden Variables: Prospects Lower bound improvement potential where I(T;V) is the mutual information u Describing correlations in a compact manner:
19
Relaxing the lower bound u Rewriting the lower bound on the log-likelihood where u The conditional entropy does not decompose The lower bound is intractable
20
Relaxing the lower bound u Using the following convexity bound u Introducing extra variational parameters u The relaxed lower bound becomes tractable Lower bound on conditional entropy
21
Optimization u Bayesian network parameters: u Smoothing parameters: u Asynchronous updates guaranties convergence
22
Results Number of time slices KL Bound
23
Discussion u Extending representational features of approximating distributions Better tradeoff ? u Addition of hidden variables improves approximation u Derivations of different methods use a uniform machinery Future directions u Saving computations by planning the order of updates u Structure of the approximating distribution
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.