8/16/99 Computer Vision: Vision and Modeling
8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling
8/16/99 A Different View of Lucas-Kanade I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2 E = ( ) = I (i) - I(i) v t i 2 i White board High Gradient has Higher weight
8/16/99 Constrained Optimization VV Constrain - I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2
8/16/99 Constraints = Subspaces E(V) VV Constrain - Analytically derived: Affine / Twist/Exponential Map Learned: Linear/non-linear Sub-Spaces
8/16/99 Motion Constraints Optical Flow: local constraints Region Layers: rigid/affine constraints Articulated: kinematic chain constraints Nonrigid: implicit / learned constraints
8/16/99 V = M ( ) Constrained Function Minimization = E(V) VV Constrain - I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2
8/16/99 2D Translation: Lucas-Kanade = E(V) VV Constrain - dx, dy... dx, dy V =V = 2D I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2
8/16/99 2D Affine: Bergen et al, Shi-Tomasi = E(V) VV Constrain - a1, a2 a3, a4 v = 6D dx dy x y i i i + I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2
8/16/99 Affine Extension Affine Motion Model: - 2D Translation - 2D Rotation - Scale in X / Y - Shear Matlab demo ->
8/16/99 Affine Extension Affine Motion Model -> Lucas-Kanade: Matlab demo ->
8/16/99 2D Affine: Bergen et al, Shi-Tomasi VV Constrain - 6D
8/16/99 K-DOF Models = E(V) VV Constrain - K-DOF V = M ( ) I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2
8/16/99 V = M ( ) Quadratic Error Norm (SSD) ??? = E(V) VV Constrain - I (1) - I(1) v t 1 I (2) - I(2) v t 2 I (n) - I(n) v t n... 2 White board (outliers?)
8/16/99 Support Maps / Layers - L2 Norm vs Robust Norm - Dangers of least square fitting: L2 D
8/16/99 Support Maps / Layers - L2 Norm vs Robust Norm - Dangers of least square fitting: L2robust DD
8/16/99 Support Maps / Layers - Robust Norm -- good for outliers - nonlinear optimization robust D
8/16/99 Support Maps / Layers - Iterative Technique Add weights to each pixel eq (white board)
8/16/99 Support Maps / Layers - how to compute weights ? -> previous iteration: how good does G-warp matches F ? -> probabilistic distance: Gaussian:
8/16/99 Error Norms / Optimization Techniques SSD: Lucas-Kanade (1981)Newton-Raphson SSD: Bergen-et al. (1992)Coarse-to-Fine SSD: Shi-Tomasi (1994)Good Features Robust Norm: Jepson-Black (1993)EM Robust Norm: Ayer-Sawhney (1995)EM + MRF MAP: Weiss-Adelson (1996)EM + MRF ML/MAP: Bregler-Malik (1998)Twists / EM ML/MAP: Irani (+Ananadan) (2000)SVD
8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling
8/16/99 Support Maps / Layers - Black-Jepson-95
8/16/99 Support Maps / Layers - More General: Layered Motion (Jepson/Black, Weiss/Adelson, …)
8/16/99 Support Maps / Layers - Special Cases of Layered Motion: - Background substraction - Outlier rejection (== robust norm) - Simplest Case: Each Layer has uniform color
8/16/99 Support Maps / Layers - Color Layers: P(skin | F(x,y))
8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling
8/16/99 Statistical Models: Represent Uncertainty and Variability Probability Theory: Proper mechanism for Uncertainty Basic Facts White Board Statistical Models / Probability Theory
8/16/99 General Performance Criteria Optimal Bayes With Applications to Classification Optimal Bayes With Applications to Classification
8/16/99 Bayes Decision Theory Example: Character Recognition: Goal: Classify new character in a way as to minimize probability of misclassification Example: Character Recognition: Goal: Classify new character in a way as to minimize probability of misclassification
8/16/99 Bayes Decision Theory 1st Concept: Priors a a b a b a a b a b a a a a b a a b a a b a a a a b b a b a b a a b a a P(a)=0.75 P(b)=0.25 ?
8/16/99 Bayes Decision Theory 2nd Concept: Conditional Probability # black pixel
8/16/99 Bayes Decision Theory Example: X=7
8/16/99 Bayes Decision Theory Example: X=8
8/16/99 Bayes Decision Theory Example: X=8 Well… P(a)=0.75 P(b)=0.25
8/16/99 Bayes Decision Theory Example: X=9 P(a)=0.75 P(b)=0.25
8/16/99 Bayes Decision Theory Bayes Theorem:
8/16/99 Bayes Decision Theory Bayes Theorem:
8/16/99 Bayes Decision Theory Bayes Theorem: Posterior = Likelihood x prior Normalization factor
8/16/99 Bayes Decision Theory Example:
8/16/99 Bayes Decision Theory Example:
8/16/99 Bayes Decision Theory Example: X>8 class b
8/16/99 Bayes Decision Theory Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries: Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries:
8/16/99 Bayes Decision Theory Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries: Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries:
8/16/99 Bayes Decision Theory Decision Regions: R1R2 R3
8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification
8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification
8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification
8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification
8/16/99 Bayes Decision Theory Discriminant functions: class membership solely based on relative sizesclass membership solely based on relative sizes Reformulate classification process in terms ofReformulate classification process in terms of discriminant functions: x is assigned to Ck if x is assigned to Ck if Discriminant functions: class membership solely based on relative sizesclass membership solely based on relative sizes Reformulate classification process in terms ofReformulate classification process in terms of discriminant functions: x is assigned to Ck if x is assigned to Ck if
8/16/99 Bayes Decision Theory Discriminant function examples:
8/16/99 Bayes Decision Theory Discriminant function examples: 2-class problem
8/16/99 Bayes Decision Theory Why is such a big deal ?
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y [/ah/, /eh/,.. /uh/] FFT melscale bank apple,...,zebra
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition FFT melscale bank /t/ /aal//aol//owl/
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition How do Humans do it?
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition “This machine can recognize speech” ??
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition “This machine can wreck a nice beach” !!
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y FFT melscale bank
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y FFT melscale bank P(“wreck a nice beach”) = P(“recognize speech”) = 0.02 Language Model
8/16/99 Bayes Decision Theory Why is such a big deal ? Example #2: Computer Vision Why is such a big deal ? Example #2: Computer Vision Low-Level Image Measurements High-Level Model Knowledge
8/16/99 Bayes Why is such a big deal ? Example #3: Curve Fitting Why is such a big deal ? Example #3: Curve Fitting E + ln p(x|c) + ln p(c)
8/16/99 Bayes Why is such a big deal ? Example #4: Snake Tracking Why is such a big deal ? Example #4: Snake Tracking E + ln p(x|c) + ln p(c)
8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling
8/16/99 Probability Density Estimation Collect Data: x1,x2,x3,x4,x5,... x x ? Estimate:
8/16/99 Probability Density Estimation Parametric Representations Non-Parametric Representations Mixture Models
8/16/99 Probability Density Estimation Parametric Representations - Normal Distribution (Gaussian) - Maximum Likelihood - Bayesian Learning
8/16/99 Normal Distribution
8/16/99 Multivariate Normal Distribution
8/16/99 Multivariate Normal Distribution Why Gaussian ? Simple analytical properties: - linear transformations of Gaussians are Gaussian - marginal and conditional densities of Gaussians are Gaussian - any moment of Gaussian densities is an explicit function of “Good” Model of Nature: - Central Limit Theorem: Mean of M random variables is distributed normally in the limit.
8/16/99 Multivariate Normal Distribution Discriminant functions:
8/16/99 Multivariate Normal Distribution Discriminant functions: equal priors + cov: Mahalanobis dist.
8/16/99 Multivariate Normal Distribution How to “learn” it from examples: Maximum Likelihood Bayesian Learning
8/16/99 Maximum Likelihood How to “learn” density from examples: x x ? ?
8/16/99 Maximum Likelihood Likelihood that density model generated data X :
8/16/99 Maximum Likelihood Likelihood that density model generated data X :
8/16/99 Maximum Likelihood Learning = optimizing (maximizing likelihood / minimizing E):
8/16/99 Maximum Likelihood Maximum Likelihood for Gaussian density: Close-form solution: