Metropolis Procedural Modeling Speaker: Alex Lin.

Slides:



Advertisements
Similar presentations
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Advertisements

Slice Sampling Radford M. Neal The Annals of Statistics (Vol. 31, No. 3, 2003)
Bayesian Estimation in MARK
CHAPTER 8 A NNEALING- T YPE A LGORITHMS Organization of chapter in ISSO –Introduction to simulated annealing –Simulated annealing algorithm Basic algorithm.
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
MDP Presentation CS594 Automated Optimal Decision Making Sohail M Yousof Advanced Artificial Intelligence.
SIGGRAPH Paper Reading 2011 Huang Haibin
Computer Vision Lab. SNU Young Ki Baik An Introduction to MCMC for Machine Learning (Markov Chain Monte Carlo)
Phylogenetic Trees Lecture 4
Markov Chains 1.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Hidden Markov Models Fundamentals and applications to bioinformatics.
Bayesian statistics – MCMC techniques
Suggested readings Historical notes Markov chains MCMC details
BAYESIAN INFERENCE Sampling techniques
CS774. Markov Random Field : Theory and Application Lecture 16 Kyomin Jung KAIST Nov
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
. Phylogeny II : Parsimony, ML, SEMPHY. Phylogenetic Tree u Topology: bifurcating Leaves - 1…N Internal nodes N+1…2N-2 leaf branch internal node.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
Strategies for Implementing Dynamic Load Sharing.
Copyright N. Friedman, M. Ninio. I. Pe’er, and T. Pupko. 2001RECOMB, April 2001 Structural EM for Phylogentic Inference Nir Friedman Computer Science &
Monte Carlo Methods in Partial Differential Equations.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Bayes Factor Based on Han and Carlin (2001, JASA).
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Phylogeny Estimation: Traditional and Bayesian Approaches Molecular Evolution, 2003
 1  Outline  stages and topics in simulation  generation of random variates.
Priors, Normal Models, Computing Posteriors
Some Probability Theory and Computational models A short overview.
1 Chapter 5 Advanced Search. 2 Chapter 5 Contents l Constraint satisfaction problems l Heuristic repair l The eight queens problem l Combinatorial optimization.
1 Gil McVean Tuesday 24 th February 2009 Markov Chain Monte Carlo.
Monte Carlo Methods1 T Special Course In Information Science II Tomas Ukkonen
Simulated Annealing.
Markov Chain Monte Carlo and Gibbs Sampling Vasileios Hatzivassiloglou University of Texas at Dallas.
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Randomized Algorithms for Bayesian Hierarchical Clustering
Molecular Systematics
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
Local Search and Optimization Presented by Collin Kanaley.
An Introduction to Markov Chain Monte Carlo Teg Grenager July 1, 2004.
Overview G. Jogesh Babu. Overview of Astrostatistics A brief description of modern astronomy & astrophysics. Many statistical concepts have their roots.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Lecture #9: Introduction to Markov Chain Monte Carlo, part 3
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Gil McVean, Department of Statistics Thursday February 12 th 2009 Monte Carlo simulation.
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Computational Physics (Lecture 10) PHY4370. Simulation Details To simulate Ising models First step is to choose a lattice. For example, we can us SC,
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Bogdan Moldovan, Ingo Thon, Jesse Davis, and Luc de Raedt
Heuristic Optimization Methods
Advanced Statistical Computing Fall 2016
Machine Learning Basics
Dynamical Statistical Shape Priors for Level Set Based Tracking
Bayesian inference Presented by Amir Hadadi
Markov chain monte carlo
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Bayesian Models in Machine Learning
Introduction to Simulated Annealing
Xin-She Yang, Nature-Inspired Optimization Algorithms, Elsevier, 2014
Presentation transcript:

Metropolis Procedural Modeling Speaker: Alex Lin

Metropolis Procedural Modeling Abstract: 1. An algorithm for controlling grammar-based procedural models 2. Optimize over the space of possible productions from the grammar 3. Specifications like geometric shapes and analytical objectives 4. Procedural models of trees, cities, buildings, and Mondrian paintings.

1. INTRODUCTION An algorithm for controlling grammar-based procedural models. Given any parametric, stochastic, conditional, context-free grammar, the algorithm takes a high-level specification of the desired model and computes a production from the grammar that matches the specification. No interaction with the grammar itself is required, and the input specification can take many forms, such as a sketch, a volumetric shape, or an analytical objective.

1. INTRODUCTION Formulate modeling operations as probabilistic inference problems over the space of productions from the grammar. Given a high-level specification of the desired model, we define an objective function that quantifies the similarity between a given production and the specification. Our goal is to optimize over the space of productions and find one that maximizes this objective.

1. INTRODUCTION Employ Markov chain Monte Carlo (MCMC) methods, which reduce optimization to sampling. Employ a generalization of traditional MCMC called Reversible jump Markov chain Monte Carlo [Green 1995], which allows the Markov chain to “jump” between spaces of varying dimension.

2. RELATED WORK Grammar-based modeling. 1. Proven effective at representing man-made structures. 2. Shape Grammar Markov chain Monte Carlo.

3. CONTEXT-FREE GRAMMARS A parametric, stochastic, conditional, context-free grammar is a tuple: V is the set of variables T is the set of terminals Σ is the set of formal parameters ω ∈ ((V ∪ T ) × R ∗ )+ is the axiom P ⊂ (V ×Σ ∗ ×B×E)×((V ∪ T)×E ∗ ) ∗ is a finite set of productions

3. CONTEXT-FREE GRAMMARS Example:

3. CONTEXT-FREE GRAMMARS Example:

3. CONTEXT-FREE GRAMMARS By searching for the optimal derivation tree instead of the optimal string, we do not need to account for this ambiguity in the probability calculations, greatly simplifying the inference process.

4. PROBABILISTIC INFERENCE FOR GRAMMARS Given a grammar G and a user-provided specification I describing a desired model, we would like to find the best derivation from G that matches the user’s input. Formulate this as a probabilistic inference problem over the space of possible derivations.

4. PROBABILISTIC INFERENCE FOR GRAMMARS 1. The described inference procedure is largely agnostic to the likelihood formulation: virtually any function can be used, provided it can be evaluated efficiently. 2. Use Markov chain Monte Carlo techniques to efficiently maximize the posterior.

5. INTRODUCTION TO MCMC INFERENCE Markov chain that takes a random, memoryless walk through a space of interest and generates a set of correlated, unbiased samples from a given function defined over that space. A Markov chain X is a sequence of random variablesX1,X2,... in a domain X. The value of the next sample in the chain depends only on the value of the sample that immediately precedes it.

5. INTRODUCTION TO MCMC INFERENCE The Metropolis-Hastings algorithm. The MH algorithm works as follows. First, X1 is set to some arbitrary point in X. To determine the value of Xi+1 at each step, a“tentative” sample x* is generated from a proposal density function q(x|xi). In theory, q can be any function that can be sampled directly and whose support includes the support of p. This tentative sample x*~q(.|xi) is either accepted into the chain as xi+1 with probability

6. REVERSIBLE JUMP MCMC The traditional Metropolis-Hastings algorithm breaks down when the dimensionality of the samples in the chain is allowed to vary. Reversible jump MCMC. RJMCMC works by supplementing the parameter-changing diffusion moves of MH with an additional set of dimension-altering jump moves, which allow the chain to move between subspaces of varying dimension.

6. REVERSIBLE JUMP MCMC 6.1 Reversibility 6.2 Dimension matching 6.3 Acceptance probability

6. REVERSIBLE JUMP MCMC 6.1 Reversibility Each move must be reversible: for any move from a model m and associated parameters xm to a model (n, xn), there must also exist a reverse move from (n, xn) back to (m, xm).

6. REVERSIBLE JUMP MCMC 6.2 Dimension matching In addition, to perform a jump from a model m to a model n, a deterministic, differentiable, invertible dimension matching function must be defined. The intuition behind them is that we are extending both models into a common, intermediary parameter space in which their densities can be meaningfully compared.

6. REVERSIBLE JUMP MCMC 6.3 Acceptance probability The RJMCMC algorithm closely resembles traditional MH. A jump is accepted into the chain with probability.

7. MCMC FOR GRAMMARS Given an grammar G along with a user-provided model specification I and associated scoring function L(I|.), we initialize the Markov chain by sampling from the prior. At every subsequent iteration, we randomly choose to apply either a jump or diffusion move in order to evolve the chain. 7.1 Diffusion moves 7.2 Jump moves

7. MCMC FOR GRAMMARS 7.1 Diffusion moves In a diffusion move, the topology of the derivation tree remains fixed, and the descriptive parameters of its derived string are modified.

7. MCMC FOR GRAMMARS 7.2 Jump moves To perform a jump on a given derivation, we first randomly select a variable v from within, and then rederive the subtree v rooted at v by re-rolling productions for v and its children. This new subtree is evolved until it consists entirely of terminal symbols or reaches the global depth limit.

7. MCMC FOR GRAMMARS 7.2 Jump moves

7. MCMC FOR GRAMMARS 7.2 Jump moves To compute the final acceptance probability for the move, we first define the jump probability of replacing the selected subtree with a new subtree.

7. MCMC FOR GRAMMARS This formula illustrates the behavior of the chain: it randomly walks through the derivation space, generally preferring to move towards models with higher likelihoods. Unlike greedy local search, however, it will occasionally accept worse moves in order to avoid becoming stuck in local maxima.

7. MCMC FOR GRAMMARS

8. OPTIMIZATION 8.1 Nonterminal selection 8.2 Parallel tempering 8.3 Delayed rejection 8.4 Annealing

8. OPTIMIZATION 8.1 Nonterminal selection While rederiving the subtree rooted at a nonterminal symbol in the lower levels of a derivation will produce only small changes in the current model, selecting a nonterminal high in the tree will cause a significant portion of the model to be replaced. To ensure that the chain proposes large and small moves in roughly equal measure, we take where is b is an estimate of the grammar’s branching factor.

8. OPTIMIZATION 8.2 Parallel tempering We employ an optimization scheme called parallel tempering [Geyer 1991]. The idea behind parallel tempering is to run a series of N independent chains at different temperatures; periodically, these chains propose to swap their states via Metropolis updates.

8. OPTIMIZATION 8.3 Delayed rejection When a jump is not accepted, we temporarily forestall the rejection and target an auxiliary diffusion move on in an effort to “clean up” the dimension-matched parameters. We then accept or reject the two moves in concert, with a specialized acceptance probability formulated to preserve ergodicity and the Markovian property of the chain [Green and Mira 1999]. In this way, the chain is able to mix well even when jump proposals generate unlikely parameters.

8. OPTIMIZATION 8.4 Annealing Use a simple annealing technique to simulate an inhomogeneous Markov chain whose stationary distribution at iteration i is, where Ti is a decreasing cooling schedule with. Under weak regularity assumptions this chain will concentrate itself on the global maxima of saving much unnecessary computation [White 1984].

9. GRAMMARS 9.1 Trees 9.2 Architecture 9.3 Mondrian art

9. GRAMMARS 8.1 Trees We developed tree grammars that simulate sympodial (as in the young oak, old oak, willow, and acacia) and monopodial (for the conifer and poplar) growth. Use a budget-based approach. Each tree starts with a budget of branches that is depleted during growth. This budget controls the probability of creating new geometry at each stage of the derivation.

9. GRAMMARS 9.1 Trees

9. GRAMMARS 9.2 Architecture Two architectural grammars: a spatial subdivision grammar for building generation, and an incremental growth grammar for city modeling.

9. GRAMMARS 9.2 Architecture

9. GRAMMARS 9.3 Mondrian art Grammar for the distinctive grid-based paintings of Piet Mondrian, a Dutch painter who co-founded the De Stijl art movement in early twentieth century. The Mondrian grammar is a probabilistic two- dimensional discrete kd-tree with bounded depth and a random color assigned to each leaf node.

10. LIKELIHOOD FORMULATIONS 10.1 Image- and volume-based modeling 10.2 Mondrian modeling

10. LIKELIHOOD FORMULATIONS 10.1 Image- and volume-based modeling

10. LIKELIHOOD FORMULATIONS 10.1 Image- and volume-based modeling

10. LIKELIHOOD FORMULATIONS 10.1 Image- and volume-based modeling

10. LIKELIHOOD FORMULATIONS 10.1 Image- and volume-based modeling

10. LIKELIHOOD FORMULATIONS 10.2 Mondrian modeling

11. PERFORMANCE

12. CONCLUSION We presented an algorithm for controlling grammar-based procedural models. The algorithm formulates procedural modeling tasks as probabilistic inference problems and solves these problems efficiently with Markov chain Monte Carlo techniques.We believe that this work takes a step towards making grammar-based procedural modeling more useful and accessible.