Bayesian Inference and Visual Processing: Image Parsing & DDMCMC. Alan Yuille (Dept. Statistics. UCLA) Tu, Chen, Yuille & Zhu (ICCV 2003).

Slides:



Advertisements
Similar presentations
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
Advertisements

1 Image Parsing: Unifying Segmentation, Detection, and Recognition Shai Bagon Oren Boiman.
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Bayesian Estimation in MARK
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
Computer Vision Lab. SNU Young Ki Baik An Introduction to MCMC for Machine Learning (Markov Chain Monte Carlo)
Qualifying Exam: Contour Grouping Vida Movahedi Supervisor: James Elder Supervisory Committee: Minas Spetsakis, Jeff Edmonds York University Summer 2009.
Bayesian Methods with Monte Carlo Markov Chains III
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Image Parsing: Unifying Segmentation and Detection Z. Tu, X. Chen, A.L. Yuille and S-C. Hz ICCV 2003 (Marr Prize) & IJCV 2005 Sanketh Shetty.
Bayesian statistics – MCMC techniques
Stochastic approximate inference Kay H. Brodersen Computational Neuroeconomics Group Department of Economics University of Zurich Machine Learning and.
Generative Models for Image Analysis Stuart Geman (with E. Borenstein, L.-B. Chang, W. Zhang)
1 Hierarchical Image-Motion Segmentation using Swendsen-Wang Cuts Adrian Barbu Siemens Corporate Research Princeton, NJ Acknowledgements: S.C. Zhu, Y.N.
Segmentation. Terminology Segmentation, grouping, perceptual organization: gathering features that belong together Fitting: associating a model with observed.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
Maximum Likelihood Network Topology Identification Mark Coates McGill University Robert Nowak Rui Castro Rice University DYNAMICS May 5 th,2003.
Prénom Nom Document Analysis: Segmentation & Layout Analysis Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Computer Vision Segmentation Marc Pollefeys COMP 256 Some slides and illustrations from D. Forsyth, T. Darrel,...
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Today Logistic Regression Decision Trees Redux Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Space-time Modelling Using Differential Equations Alan E. Gelfand, ISDS, Duke University (with J. Duan and G. Puggioni)
Modeling Vision as Bayesian Inference: Is it Worth the Effort? Alan L. Yuille. UCLA. Co-Director: Centre for Image and Vision Sciences. Dept. Statistics.
Vision in Man and Machine. STATS 19 SEM Talk 2. Alan L. Yuille. UCLA. Dept. Statistics and Psychology.
Computer Vision Segmentation Marc Pollefeys COMP 256 Some slides and illustrations from D. Forsyth, T. Darrel,...
Data-Driven Markov Chain Monte Carlo Presented by Tomasz MalisiewiczTomasz Malisiewicz for Advanced PerceptionAdvanced Perception 3/1/2006.
Grammar of Image Zhaoyin Jia, Problems  Enormous amount of vision knowledge:  Computational complexity  Semantic gap …… Classification,
Computer Vision Systems for the Blind and Visually Disabled. STATS 19 SEM Talk 3. Alan Yuille. UCLA. Dept. Statistics and Psychology.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Computer vision: models, learning and inference
Bayes Factor Based on Han and Carlin (2001, JASA).
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
WSEAS AIKED, Cambridge, Feature Importance in Bayesian Assessment of Newborn Brain Maturity from EEG Livia Jakaite, Vitaly Schetinin and Carsten.
Image Segmentation Seminar III Xiaofeng Fan. Today ’ s Presentation Problem Definition Problem Definition Approach Approach Segmentation Methods Segmentation.
Leo Zhu CSAIL MIT Joint work with Chen, Yuille, Freeman and Torralba 1.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Graphical Models in Vision. Alan L. Yuille. UCLA. Dept. Statistics.
Supervised Learning of Edges and Object Boundaries Piotr Dollár Zhuowen Tu Serge Belongie.
Markov Random Fields Probabilistic Models for Images
An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Randomized Algorithms for Bayesian Hierarchical Clustering
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 1. Stat 231. A.L. Yuille. Fall 2004 AdaBoost.. Binary Classification. Read 9.5 Duda,
Parsing Images with Context/Content Sensitive Grammars Eran Borenstein, Stuart Geman, Ya Jin, Wei Zhang.
Tracking Multiple Cells By Correspondence Resolution In A Sequential Bayesian Framework Nilanjan Ray Gang Dong Scott T. Acton C.L. Brown Department of.
Sequential Monte-Carlo Method -Introduction, implementation and application Fan, Xin
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Unsupervised Mining of Statistical Temporal Structures in Video Liu ze yuan May 15,2011.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
Ch.8 Efficient Coding of Visual Scenes by Grouping and Segmentation Bayesian Brain Tai Sing Lee and Alan L. Yuille Heo, Min-Oh.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Segmentation Through Optimization Pyry Matikainen.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Machine Vision ENT 273 Lecture 4 Hema C.R.
MCMC Output & Metropolis-Hastings Algorithm Part I
Dynamical Statistical Shape Priors for Level Set Based Tracking
Data-Driven Markov Chain Monte Carlo
Outline Image Segmentation by Data-Driven Markov Chain Monte Carlo
Edges/curves /blobs Grammars are important because:
Image Parsing & DDMCMC. Alan Yuille (Dept. Statistics. UCLA)
Robust Full Bayesian Learning for Neural Networks
Expectation-Maximization & Belief Propagation
Presentation transcript:

Bayesian Inference and Visual Processing: Image Parsing & DDMCMC. Alan Yuille (Dept. Statistics. UCLA) Tu, Chen, Yuille & Zhu (ICCV 2003).

Something Completely Different? No Medical Images. No Medical Images. Segmentation and Recognition of Natural Images. Segmentation and Recognition of Natural Images. But the ideas can be applied to Medical Images… But the ideas can be applied to Medical Images…

Image Parsing. (I) Image are composed of visual patterns: (I) Image are composed of visual patterns: (II) Parse an image by decomposing it into patterns. (II) Parse an image by decomposing it into patterns.

Talk Part I: Generative Model. Stochastic grammar for generating images in terms of visual patterns. Stochastic grammar for generating images in terms of visual patterns. Visual patterns can be generic (texture/shading) or objects. Visual patterns can be generic (texture/shading) or objects. Hierarchical Probability model – probability on graphs. Hierarchical Probability model – probability on graphs.

Talk Part II: Inference Algorithm Interpreting an image corresponds to constructing a parse graph. Interpreting an image corresponds to constructing a parse graph. Set of moves for constructing the parse graph. Set of moves for constructing the parse graph. Dynamics for moves use bottom-up & top-down visual processing. Dynamics for moves use bottom-up & top-down visual processing. Data-Driven Markov Chain Monte Carlo (DDMCMC). Data-Driven Markov Chain Monte Carlo (DDMCMC). Discriminative Models to drive Generative models. Discriminative Models to drive Generative models.

Part I: Generative Models. Previous related work by our group: Previous related work by our group: Zhu & Yuille 1996 (Region Competition). Zhu & Yuille 1996 (Region Competition). Tu & Zhu Tu & Zhu Tu & Zhu Tu & Zhu These theories assumed generic visual patterns only. These theories assumed generic visual patterns only.

Generic Patterns & Object Patterns. Limitations of Generic Visual Patterns. Limitations of Generic Visual Patterns. Object patterns enable us to unify segmentation & recognition. Object patterns enable us to unify segmentation & recognition.

Stochastic Grammar: Parsing Graph. Nodes represent visual patterns. Child nodes to image pixels. Nodes represent visual patterns. Child nodes to image pixels. Stochastic Grammars: Manning & Schultz.

Image Patterns. Node attributes: Node attributes: Zeta: Pattern Type – 66 Zeta: Pattern Type – 66 (I) Smooth, (II) Texture/Cluster, (III) Shading. (IV) Face, (V– LXVI) Text Characters. (I) Smooth, (II) Texture/Cluster, (III) Shading. (IV) Face, (V– LXVI) Text Characters. L – shape descriptor (image region modeled). L – shape descriptor (image region modeled). Theta: Model parameters. Theta: Model parameters.

Generative Model: Likelihood: Likelihood: Prior: Prior: Samples: Samples:

Stochastic Grammar Summary. Graph represents: Graph represents: Sampling from the graph generates an image. Sampling from the graph generates an image.

Part II: Inference Algorithm. We described a model to generate image: We described a model to generate image: P(I|W) & P(W). P(I|W) & P(W). Need an algorithm to infer W* from P(I|W) & P(W). Need an algorithm to infer W* from P(I|W) & P(W).

Inference & Parse Graph. Inference requires constructing a parse graph. Inference requires constructing a parse graph. Dynamics to create/delete nodes and alter node attributes: Dynamics to create/delete nodes and alter node attributes:

Moves: Birth & Death of Text. Birth & Death of Text. Birth & Death of Faces. Birth & Death of Faces. Splitting & Merging of Regions. Splitting & Merging of Regions. Switching Node Attributes (Model Parameters & Pattern Types). Switching Node Attributes (Model Parameters & Pattern Types). Moving Region Boundaries. Moving Region Boundaries.

Markov Chain Monte Carlo. Design a Markov Chain (MC) with transition kernel Design a Markov Chain (MC) with transition kernel Satisfies Detailed Balance. Satisfies Detailed Balance. Then repeated sampling from the MC will converge to samples from the posterior P(W|I). Then repeated sampling from the MC will converge to samples from the posterior P(W|I).

Moves & Sub-kernels. Implement each move by a transition sub- kernel: Implement each move by a transition sub- kernel: Combines moves by a full kernel: Combines moves by a full kernel: At each time-step – choose a type of move, then apply it to the graph. At each time-step – choose a type of move, then apply it to the graph. Kernels obey: Kernels obey:

Data Driven Proposals. Use data-driven proposals to make the Markov Chain efficient. Use data-driven proposals to make the Markov Chain efficient. Metropolis-Hastings design: Metropolis-Hastings design: Proposal probabilities are based on discriminative cues. Proposal probabilities are based on discriminative cues.

Discriminative Methods: Edge Cues Edge Cues Binarization Cues. Binarization Cues. Face Region Cues (AdaBoost). Face Region Cues (AdaBoost). Text Region Cues (AdaBoost). Text Region Cues (AdaBoost). Shape Affinity Cues. Shape Affinity Cues. Region Affinity Cues. Region Affinity Cues. Model Parameters & Pattern Type. Model Parameters & Pattern Type.

Design Proposals I. How to generate proposals How to generate proposals is the scope of W– states that can be reached from W with one move of type i. is the scope of W– states that can be reached from W with one move of type i. Ideal proposal: Ideal proposal:

Design Proposals II. Re-express this as: Re-express this as: Set Q(W,W’|I) to approximate P(W’|I)/P(W|I) and be easily computatable. Set Q(W,W’|I) to approximate P(W’|I)/P(W|I) and be easily computatable.

Example: Create Node by Splitting Select region (node) R_k. Select region (node) R_k. Propose a finite set of ways to split R_k based on discriminative cues (edges and edge linking). Propose a finite set of ways to split R_k based on discriminative cues (edges and edge linking). Consider split R_k to R_i & R_j. Consider split R_k to R_i & R_j.

Example: Create Node by Splitting. Create (Split). Create (Split). Denominator is known, we are in state W. Denominator is known, we are in state W. Use an affinity measure Use an affinity measure

Example: Create Face Create Face. Create Face. Bottom-up proposal for face driven by AdaBoost & edge detection. Bottom-up proposal for face driven by AdaBoost & edge detection. This will require splitting an existing region R_k into R_i (face) and R_j (remainder). This will require splitting an existing region R_k into R_i (face) and R_j (remainder). Parameters for R_k & R_j are known (the same). Parameters for R_k & R_j are known (the same).

Examples: Death by Merging. Death (Merge). Death (Merge). Select regions R_i & R_j to merge based on a similarity measure (as for splitting). Select regions R_i & R_j to merge based on a similarity measure (as for splitting). Same approximations as for splitting to approximate. Same approximations as for splitting to approximate.

Node Splitting/Merging. Causes for split/merge. Causes for split/merge. (i) Bottom-up cue – there is probably a face or text here (AdaBoost + Edge Detection). (i) Bottom-up cue – there is probably a face or text here (AdaBoost + Edge Detection). (ii) Bottom-up cue – there is an intensity edge splitting the region. (ii) Bottom-up cue – there is an intensity edge splitting the region. (iii) Current state W – model for region I fits data poorly. (iii) Current state W – model for region I fits data poorly. (iv) Current state W -- two regions are similar (by affinity measure). (iv) Current state W -- two regions are similar (by affinity measure).

Full Strategy: Integration: Integration:

Control Strategy: (For AZ).

AdaBoost – Conditional Probs. Supervised Learning. Supervised Learning.

Experiments: Competition & Cooperation. Competition & Cooperation.

Experiments: More results: More results:

Key Ideas: Generative Models for Visual Patterns & Stochastic Grammars. Generative Models for Visual Patterns & Stochastic Grammars. Inference: set of “moves” on the parse graph implemented by Kernels. Inference: set of “moves” on the parse graph implemented by Kernels. Discriminative Models – bottom-up – drive top-down Generative Models. Discriminative Models – bottom-up – drive top-down Generative Models.

Discussion: Image parsing gives a way to combine segmentation with recognition. Image parsing gives a way to combine segmentation with recognition. The DDMCMC algorithm is a way to combine discriminative models with generative models. The DDMCMC algorithm is a way to combine discriminative models with generative models. This bottom-up/top-down architecture may be suggestive of brain architecture. This bottom-up/top-down architecture may be suggestive of brain architecture.