Siddharth Choudhary.  Refines a visual reconstruction to produce jointly optimal 3D structure and viewing parameters  ‘bundle’ refers to the bundle.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Instabilities of SVD Small eigenvalues -> m+ sensitive to small amounts of noise Small eigenvalues maybe indistinguishable from 0 Possible to remove small.
Factorial Mixture of Gaussians and the Marginal Independence Model Ricardo Silva Joint work-in-progress with Zoubin Ghahramani.
Online Social Networks and Media. Graph partitioning The general problem – Input: a graph G=(V,E) edge (u,v) denotes similarity between u and v weighted.
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
A Short Introduction to Curve Fitting and Regression by Brad Morantz
The loss function, the normal equation,
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Methods For Nonlinear Least-Square Problems
Robust Estimator 學生 : 范育瑋 老師 : 王聖智. Outline Introduction LS-Least Squares LMS-Least Median Squares RANSAC- Random Sample Consequence MLESAC-Maximum likelihood.
Avoiding Communication in Sparse Iterative Solvers Erin Carson Nick Knight CS294, Fall 2011.
Data Modeling and Least Squares Fitting 2 COS 323.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Expectation Maximization Algorithm
Nonlinear Regression Probability and Statistics Boris Gervits.
Optimization Methods One-Dimensional Unconstrained Optimization
Improved BP algorithms ( first order gradient method) 1.BP with momentum 2.Delta- bar- delta 3.Decoupled momentum 4.RProp 5.Adaptive BP 6.Trinary BP 7.BP.
12 1 Variations on Backpropagation Variations Heuristic Modifications –Momentum –Variable Learning Rate Standard Numerical Optimization –Conjugate.
Advanced Topics in Optimization
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Optimization Methods One-Dimensional Unconstrained Optimization
Lecture 12: Structure from motion CS6670: Computer Vision Noah Snavely.
CSC2535: 2013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
Collaborative Filtering Matrix Factorization Approach
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
Last Lecture (optical center) origin principal point P (X,Y,Z) p (x,y) x y.
Chem Math 252 Chapter 5 Regression. Linear & Nonlinear Regression Linear regression –Linear in the parameters –Does not have to be linear in the.
Camera calibration Digital Visual Effects Yung-Yu Chuang with slides by Richard Szeliski, Steve Seitz,, Fred Pighin and Marc Pollefyes.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Solution of Sparse Linear Systems
Direct Methods for Sparse Linear Systems Lecture 4 Alessandra Nardi Thanks to Prof. Jacob White, Suvranu De, Deepak Ramaswamy, Michal Rewienski, and Karen.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Bundle Adjustment A Modern Synthesis Bill Triggs, Philip McLauchlan, Richard Hartley and Andrew Fitzgibbon Presentation by Marios Xanthidis 5 th of No.
Variations on Backpropagation.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
CS 290H Lecture 15 GESP concluded Final presentations for survey projects next Tue and Thu 20-minute talk with at least 5 min for questions and discussion.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Numerical Analysis – Data Fitting Hanyang University Jong-Il Park.
Symmetric-pattern multifrontal factorization T(A) G(A)
Computational Intelligence: Methods and Applications Lecture 22 Linear discrimination - variants Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Conjugate gradient iteration One matrix-vector multiplication per iteration Two vector dot products per iteration Four n-vectors of working storage x 0.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Probability Theory and Parameter Estimation I
Solving Linear Systems Ax=b
Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/4/17
Digital Visual Effects, Spring 2006 Yung-Yu Chuang 2005/4/19
Parameter estimation class 5
Going Backwards In The Procedure and Recapitulation of System Identification By Ali Pekcan 65570B.
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Non-linear Least-Squares
Collaborative Filtering Matrix Factorization Approach
Digital Visual Effects Yung-Yu Chuang
Variations on Backpropagation.
Structure from Motion with Non-linear Least Squares
Nonlinear regression.
Noah Snavely.
CS 290H Lecture 3 Fill: bounds and heuristics
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
Variations on Backpropagation.
Structure from Motion with Non-linear Least Squares
Parameter estimation class 6
Presentation transcript:

Siddharth Choudhary

 Refines a visual reconstruction to produce jointly optimal 3D structure and viewing parameters  ‘bundle’ refers to the bundle of light rays leaving each 3D feature and converging on each camera center.

 Minimization of weighted sum of squared error ( SSE ) cost function:

 Least-squares fitting is a maximum likelihood estimation of the fitted parameters if the measurement errors are independent and normally distributed with constant standard deviation  The probability distribution of the sum of a very large number of very small random deviations almost always converges to a normal distribution.

 It is highly sensitive to outliers, because the Gaussian has extremely small tails compared to most real measurement error distribution. ( It is the reason of using Hierarchical SFM ) Gaussian Tail problem and its effects is addressed in the paper ‘ Pushing the envelope of modern bundle adjustment techniques, CVPR 2010’

 Gradient Descent Method  Newton-Rhapson Method  Gauss – Newton Method  Levenberg – Marquardt Method

 A first-order optimization algorithm.  To find a local minimum of a function using gradient descent, one takes steps proportional to the negative of the gradient of the function at the current point. While k<k max

 It is robust when x is far from optimum but has poor final convergence ( this fact is used in designing the LM iteration )

 It is a second order optimization method  Newton's method can often converge remarkably quickly, especially if the iteration begins "sufficiently near" the desired root.

 The Gauss–Newton algorithm is a method used to solve non-linear least squares problems

 For well-parametrized bundle problems under an outlier-free least squares cost model evaluated near the cost minimum, the Gauss- Newton approximation is usually very accurate

 The LMA interpolates between the Gauss– Newton algorithm (GNA) and the method of gradient descent.  When far from the minimum it acts as a steepest descent and it performs gauss newton iteration when near to the solution.

 Second order optimization methods like Gauss – Newton and LM requires a few but heavy iterations  First order optimization methods like Gradient descent requires a lot of light iterations.

 Exploit the problem structure  Use factorization effectively  Use stable local parametrizations  Scaling and preconditioning

 The Schur Complement and the reduced camera system  Cholesky Decomposition  Sparse Factorization  Variable Ordering ▪ Top down ordering ▪ Bottom up ordering  Preconditioning  Conjugate Gradient method  Multigrid Methods

Left Multiplyto both sides Reduced Camera System

 Since both the Hessian and the reduced camera system is sparse for large scale systems, sparse factorization methods are preferred.  Variable Ordering  Preconditioning  Conjugate Gradient Method  Parallel Multigrid Methods

 There is a phenomenon of fill – in.  After each step, we have more number of non – zeros which lead to more number of floating point operations.

 The effect of cholesky factorization after variables are re ordered creates the least fill- in  The task of variable ordering is to reorder the matrix to create the least fill in.

Finding the ordering which results in the least fill- in is a NP-complete problem Some of the heuristics used are:  Minimum Degree Reordering ( Bottom – up approach )  Nested Dissection ( Top – Down approach )  These methods gives an idea of sparsity and structure of matrices.

 Neighbors of eliminated vertex in previous graph become clique (fully connected subgraph) in modified graph.  Entries of A that were initially zero, may become non zero entries, called fill

 Since finding the order of vertices with minimum fill in is a NP – Complete problem  This is a greedy algorithm such that after each iteration we select a vertex with minimum degree.  This is a bottom up method trying to minimize fill-in locally and greedily at each step, at the risk of global short sightedness

 Form the Elimination Graph.  Recursively partition the graph into subgraphs using separators, small subsets of vertices the removal of which allows the graph to be partitioned into subgraphs with at most a constant fraction of the number of vertices.  Perform Cholesky decomposition (a variant of Gaussian elimination for symmetric matrices), ordering the elimination of the variables by the recursive structure of the partition: each of the two subgraphs formed by removing the separator is eliminated first, and then the separator vertices are eliminated.

 It is an iterative method to solve a sparse system large enough to be handled by Cholesky decomposition  Converges in at most n steps where n is the size of the matrix

Thank You