Image & Model Fitting Abstractions February 2017

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Scaling Up Graphical Model Inference
Linear Time Methods for Propagating Beliefs Min Convolution, Distance Transforms and Box Sums Daniel Huttenlocher Computer Science Department December,
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
State Estimation and Kalman Filtering CS B659 Spring 2013 Kris Hauser.
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Exact Inference in Bayes Nets
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Supervised Learning Recap
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Computer vision: models, learning and inference
Belief Propagation by Jakob Metzler. Outline Motivation Pearl’s BP Algorithm Turbo Codes Generalized Belief Propagation Free Energies.
Aspects of Conditional Simulation and estimation of hydraulic conductivity in coastal aquifers" Luit Jan Slooten.
Hidden Markov Models Theory By Johan Walters (SR 2003)
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Visual Recognition Tutorial
Machine Learning Neural Networks
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Motion Analysis (contd.) Slides are from RPI Registration Class.
CSci 6971: Image Registration Lecture 4: First Examples January 23, 2004 Prof. Chuck Stewart, RPI Dr. Luis Ibanez, Kitware Prof. Chuck Stewart, RPI Dr.
Lecture 5: Learning models using EM
1 Markov random field: A brief introduction Tzu-Cheng Jen Institute of Electronics, NCTU
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
כמה מהתעשייה? מבנה הקורס השתנה Computer vision.
Collaborative Filtering Matrix Factorization Approach
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by B.-H. Kim Biointelligence Laboratory, Seoul National.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Artificial Neural Networks
Markov Random Fields Probabilistic Models for Images
Problems in large-scale computer vision David Crandall School of Informatics and Computing Indiana University.
Forward-Scan Sonar Tomographic Reconstruction PHD Filter Multiple Target Tracking Bayesian Multiple Target Tracking in Forward Scan Sonar.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
CS Statistical Machine learning Lecture 24
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Lecture 2: Statistical learning primer for biologists
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Final Exam Review CS479/679 Pattern Recognition Dr. George Bebis 1.
Markov Random Fields & Conditional Random Fields
Pattern Recognition and Machine Learning
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Optimization Indiana University July Geoffrey Fox
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Overview G. Jogesh Babu. R Programming environment Introduction to R programming language R is an integrated suite of software facilities for data manipulation,
Geoffrey Fox Panel Talk: February
Optimization: Algorithms and Applications
Introduction of BP & TRW-S
Pathology Spatial Analysis February 2017
Implementing parts of HPC-ABDS in a multi-disciplinary collaboration
10701 / Machine Learning.
Data Mining Lecture 11.
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Tutorial Overview February 2017
Markov Networks.
Probabilistic Models with Latent Variables
Collaborative Filtering Matrix Factorization Approach
CSCI 5822 Probabilistic Models of Human and Machine Learning
Scalable Parallel Interoperable Data Analytics Library
Integration and Graphical Models
An introduction to Graphical Models – Michael Jordan
Expectation-Maximization & Belief Propagation
Boltzmann Machine (BM) (§6.4)
Indiana University July Geoffrey Fox
Markov Networks.
Presentation transcript:

Image & Model Fitting Abstractions February 2017 NSF 1443054: CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science Software: MIDAS HPC-ABDS Image & Model Fitting Abstractions February 2017

Imaging applications Many scientific domains now collect large scale image data, e.g. Astronomy: wide-area telescope data Ecology, meteorology: Satellite imagery Biology, neuroscience: Live-cell imaging, MRIs, … Medicine: X-ray, MRI, CT, … Physics, chemistry: electron microscopy, … Earth science: Sonar, satellite, radar, … Challenge has moved from collecting data to analyzing it Large scale (number of images or size of images) overwhelming for human analysis Recent progress in computer vision makes reliable automated image analysis feasible

Key image analysis problems Many names for similar problems; most fall into: Segmentation: Dividing image into homogeneous regions Detection, recognition: Finding and identifying important structures and their properties Reconstruction: Inferring properties of a data source from noisy, incomplete observations (e.g. removing noise from an image, estimating 3d structure of scene from multiple images) Matching and alignment: Finding correspondences between images Most of these problems can be thought of as image pre-processing followed by model fitting Arbelaez 2011 Dollar 2012 Crandall 2013

SPIDAL image abstractions SPIDAL has or will have support for imaging at several levels of abstractions: Low-level: image processing (e.g. filtering, denoising), local/global feature extraction Mid-level: object detection, image segmentation, object matching, 3D feature extraction, image registration Application level: radar informatics, polar image analysis, spatial image analysis, pathology image analysis

SPIDAL model-fitting abstractions Most image analysis relies on some form of model fitting: Segmentation: fitting parameterized regions (e.g. contiguous regions) to an image Object detection: fitting object model to an image Registration and alignment: fitting model of image transformation (e.g. warping) between multiple images Reconstruction: fitting prior information about the visual world to observed data Usually high degree of noise and outliers, so not a simple matter of e.g. linear regression or constraint satisfaction! Instead involves defining an energy function or error function, and finding minima of that error function

SPIDAL model-fitting abstractions SPIDAL has or will have support for model fitting at several levels of abstractions: Low-level: grid search, Viterbi, Forward-Backward, Markov Chain Monte Carlo (MCMC) algorithms, deterministic simulated annealing, gradient descent Mid-level: Support Vector Machine learning, Random Forest learning, K-means, vector clustering, Latent Dirichlet Allocation Application level: Spatial clustering, image clustering

Energy minimization (optimization) Very general idea: find parameters of a model that minimize an energy (or cost function), given a set of data Global minima easy to find if energy function is simple (e.g. convex) Energy function usually has unknown number & distribution of local minima; global minimum very difficult to find Many algorithms tailored to cost functions for specific applications, usually some heuristics to encourage finding “good” solutions, rarely theoretical guarantees. High computation cost. Remember deterministic annealing - Arman Bahl

Common energy minimization cases Parameter space: Continuous vs. Discrete Energy functions with particular forms, e.g.: Hidden Markov Model: chain of observable and unobservable variables. Each unknown variable is a (nondeterministic) function of its observable variable, and the two unobservables before and after. Markov Random Field: generalization of HMM, each unobservable variable is function of a small number of neighboring unobservables.

Continuous optimization Most techniques rely on gradient descent, “hill-climbing” E.g. Newton’s method with various heuristics to escape local minima Support in SPIDAL Levenberg-Marquardt Deterministic annealing Custom methods as in neural networks or SMACOF for MDS

Discrete optimization support in SPIDAL Grid search: trivially parallelizable but inefficient Viterbi and Forward-Backward: efficient exact algorithms for Maximum A Posteriori (MAP) and marginal inference using dynamic programming, but restricted to Hidden Markov Models. Loopy Belief Propagation: approximate algorithm for MAP inference on Markov Random Field models. No optimality or even convergence guarantees, but applicable to a general class of models. Tree ReWeighted Message Passing (TRW): approximate algorithm for MAP inference on some MRFs. Computes bounds that often give meaningful measure of quality of solution (with respect to unknown global minimum). Markov Chain Monte Carlo: approximate algorithms for graphical models including HMMs, MRFs, and Bayes Nets in general.

SPIDAL Algorithms – Optimization I Manxcat: Levenberg Marquardt Algorithm for non-linear 2 optimization with sophisticated version of Newton’s method calculating value and derivatives of objective function. Parallelism in calculation of objective function and in parameters to be determined. Complete – needs SPIDAL Java optimization Viterbi algorithm, for finding the maximum a posteriori (MAP) solution for a Hidden Markov Model (HMM). The running time is O(n*s^2) where n is the number of variables and s is the number of possible states each variable can take. We will provide an "embarrassingly parallel" version that processes multiple problems (e.g. many images) independently; parallelizing within the same problem not needed in our application space. Needs Packaging in SPIDAL Forward-backward algorithm, for computing marginal distributions over HMM variables. Similar characteristics as Viterbi above. Needs Packaging in SPIDAL

SPIDAL Algorithms – Optimization II Loopy belief propagation (LBP) for approximately finding the maximum a posteriori (MAP) solution for a Markov Random Field (MRF). Here the running time is O(n^2*s^2*i) in the worst case where n is number of variables, s is number of states per variable, and i is number of iterations required (which is usually a function of n, e.g. log(n) or sqrt(n)). Here there are various parallelization strategies depending on values of s and n for any given problem. We will provide two parallel versions: embarrassingly parallel version for when s and n are relatively modest, and parallelizing each iteration of the same problem for common situation when s and n are quite large so that each iteration takes a long time relative to number of iterations required. Needs Packaging in SPIDAL Markov Chain Monte Carlo (MCMC) for approximately computing distributions and sampling over MRF variables. Similar to LBP with the same two parallelization strategies. Needs Packaging in SPIDAL

Higher-level model fitting Clustering: K-means, vector clustering Topic modeling: Latent Dirichlet Allocation Machine learning: Random Forests, Support Vector Machines Applications: spatial clustering, image clustering Plate notation for smoothed LDA Random Forest

Two exemplar applications: Polar science and Pathology imaging Despite very different applications, data, and approaches, same key abstractions apply! Segmentation: divide radar imagery into ice vs rock, or pathology images into parts of cells, etc. Recognition: subsurface features of ice, organism components in biology Reconstruction: estimate 3d structure of ice, or 3d structure of organisms