8/16/99 Computer Vision: Vision and Modeling. 8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction,

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pattern Classification & Decision Theory. How are we doing on the pass sequence? Bayesian regression and estimation enables us to track the man in the.
Pattern Recognition and Machine Learning
Face Recognition Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
An Introduction of Support Vector Machine
Chapter 2: Bayesian Decision Theory (Part 2) Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density.
Pattern Classification, Chapter 2 (Part 2) 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R.
Pattern Classification. Chapter 2 (Part 1): Bayesian Decision Theory (Sections ) Introduction Bayesian Decision Theory–Continuous Features.
Pattern Classification Chapter 2 (Part 2)0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O.
Chapter 4: Linear Models for Classification
Computer vision: models, learning and inference
Visual Recognition Tutorial
Bayesian Decision Theory Chapter 2 (Duda et al.) – Sections
MASKS © 2004 Invitation to 3D vision Lecture 8 Segmentation of Dynamical Scenes.
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Announcements  Homework 4 is due on this Thursday (02/27/2004)  Project proposal is due on 03/02.
A gentle introduction to Gaussian distribution. Review Random variable Coin flip experiment X = 0X = 1 X: Random variable.
Project 4 out today –help session today –photo session today Project 2 winners Announcements.
Bayesian Frameworks for Deformable Pattern Classification and Retrieval by Kwok-Wai Cheung January 1999.
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation Given.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
Principles of Pattern Recognition
Lecture 2: Bayesian Decision Theory 1. Diagram and formulation
Learning and Recognizing Human Dynamics in Video Sequences Christoph Bregler Alvina Goh Reading group: 07/06/06.
ECE 8443 – Pattern Recognition LECTURE 03: GAUSSIAN CLASSIFIERS Objectives: Normal Distributions Whitening Transformations Linear Discriminants Resources.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
CS Statistical Machine learning Lecture 10 Yuan (Alan) Qi Purdue CS Sept
1 E. Fatemizadeh Statistical Pattern Recognition.
2/14/00 Computer Vision. 2/14/00 Computer Vision Lecturer: Ir. Resmana Lim, M.Eng. Text: 1) Computer Vision -- A Modern Approach.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
BCS547 Neural Decoding.
Linear Models for Classification
Lecture 2: Statistical learning primer for biologists
Final Review Course web page: vision.cis.udel.edu/~cv May 21, 2003  Lecture 37.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Probability Theory and Parameter Estimation I
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Ch3: Model Building through Regression
LECTURE 03: DECISION SURFACES
Segmentation of Dynamic Scenes
Statistical Models for Automatic Speech Recognition
Special Topics In Scientific Computing
Data Mining Lecture 11.
Lecture 26: Faces and probabilities
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
CS4670: Intro to Computer Vision
Pattern Recognition and Machine Learning
Segmentation of Dynamical Scenes
Generally Discriminant Analysis
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Announcements Project 2 artifacts Project 3 due Thursday night
Announcements Project 4 out today Project 2 winners help session today
Announcements Artifact due Thursday
EM Algorithm and its Applications
Announcements Artifact due Thursday
The “Margaret Thatcher Illusion”, by Peter Thompson
Presentation transcript:

8/16/99 Computer Vision: Vision and Modeling

8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling

8/16/99 A Different View of Lucas-Kanade I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2    E =  ( ) = I (i) -  I(i) v t i  2 i  White board High Gradient has Higher weight

8/16/99 Constrained Optimization VV Constrain - I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2   

8/16/99 Constraints = Subspaces E(V) VV Constrain - Analytically derived: Affine / Twist/Exponential Map Learned: Linear/non-linear Sub-Spaces

8/16/99 Motion Constraints Optical Flow: local constraints Region Layers: rigid/affine constraints Articulated: kinematic chain constraints Nonrigid: implicit / learned constraints

8/16/99 V = M (   ) Constrained Function Minimization = E(V) VV Constrain - I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2   

8/16/99 2D Translation: Lucas-Kanade = E(V) VV Constrain - dx, dy... dx, dy V =V = 2D I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2   

8/16/99 2D Affine: Bergen et al, Shi-Tomasi = E(V) VV Constrain - a1, a2 a3, a4 v = 6D dx dy x y i i i + I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2   

8/16/99 Affine Extension Affine Motion Model: - 2D Translation - 2D Rotation - Scale in X / Y - Shear Matlab demo ->

8/16/99 Affine Extension Affine Motion Model -> Lucas-Kanade: Matlab demo ->

8/16/99 2D Affine: Bergen et al, Shi-Tomasi VV Constrain - 6D

8/16/99 K-DOF Models = E(V) VV Constrain - K-DOF V = M (   ) I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2   

8/16/99 V = M (   ) Quadratic Error Norm (SSD) ??? = E(V) VV Constrain - I (1) -  I(1) v t 1 I (2) -  I(2) v t 2 I (n) -  I(n) v t n... 2     White board (outliers?)

8/16/99 Support Maps / Layers - L2 Norm vs Robust Norm - Dangers of least square fitting: L2 D

8/16/99 Support Maps / Layers - L2 Norm vs Robust Norm - Dangers of least square fitting: L2robust DD

8/16/99 Support Maps / Layers - Robust Norm -- good for outliers - nonlinear optimization robust D

8/16/99 Support Maps / Layers - Iterative Technique Add weights to each pixel eq (white board)

8/16/99 Support Maps / Layers - how to compute weights ? -> previous iteration: how good does G-warp matches F ? -> probabilistic distance: Gaussian:

8/16/99 Error Norms / Optimization Techniques SSD: Lucas-Kanade (1981)Newton-Raphson SSD: Bergen-et al. (1992)Coarse-to-Fine SSD: Shi-Tomasi (1994)Good Features Robust Norm: Jepson-Black (1993)EM Robust Norm: Ayer-Sawhney (1995)EM + MRF MAP: Weiss-Adelson (1996)EM + MRF ML/MAP: Bregler-Malik (1998)Twists / EM ML/MAP: Irani (+Ananadan) (2000)SVD

8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling

8/16/99 Support Maps / Layers - Black-Jepson-95

8/16/99 Support Maps / Layers - More General: Layered Motion (Jepson/Black, Weiss/Adelson, …)

8/16/99 Support Maps / Layers - Special Cases of Layered Motion: - Background substraction - Outlier rejection (== robust norm) - Simplest Case: Each Layer has uniform color

8/16/99 Support Maps / Layers - Color Layers: P(skin | F(x,y))

8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling

8/16/99 Statistical Models: Represent Uncertainty and Variability Probability Theory: Proper mechanism for Uncertainty Basic Facts  White Board Statistical Models / Probability Theory

8/16/99 General Performance Criteria Optimal Bayes With Applications to Classification Optimal Bayes With Applications to Classification

8/16/99 Bayes Decision Theory Example: Character Recognition: Goal: Classify new character in a way as to minimize probability of misclassification Example: Character Recognition: Goal: Classify new character in a way as to minimize probability of misclassification

8/16/99 Bayes Decision Theory 1st Concept: Priors a a b a b a a b a b a a a a b a a b a a b a a a a b b a b a b a a b a a P(a)=0.75 P(b)=0.25 ?

8/16/99 Bayes Decision Theory 2nd Concept: Conditional Probability # black pixel

8/16/99 Bayes Decision Theory Example: X=7

8/16/99 Bayes Decision Theory Example: X=8

8/16/99 Bayes Decision Theory Example: X=8 Well… P(a)=0.75 P(b)=0.25

8/16/99 Bayes Decision Theory Example: X=9 P(a)=0.75 P(b)=0.25

8/16/99 Bayes Decision Theory Bayes Theorem:

8/16/99 Bayes Decision Theory Bayes Theorem:

8/16/99 Bayes Decision Theory Bayes Theorem: Posterior = Likelihood x prior Normalization factor

8/16/99 Bayes Decision Theory Example:

8/16/99 Bayes Decision Theory Example:

8/16/99 Bayes Decision Theory Example: X>8 class b

8/16/99 Bayes Decision Theory Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries: Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries:

8/16/99 Bayes Decision Theory Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries: Goal: Classify new character in a way as to minimize probability of misclassification Decision boundaries:

8/16/99 Bayes Decision Theory Decision Regions: R1R2 R3

8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification

8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification

8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification

8/16/99 Bayes Decision Theory Goal: minimize probability of misclassification

8/16/99 Bayes Decision Theory Discriminant functions: class membership solely based on relative sizesclass membership solely based on relative sizes Reformulate classification process in terms ofReformulate classification process in terms of discriminant functions: x is assigned to Ck if x is assigned to Ck if Discriminant functions: class membership solely based on relative sizesclass membership solely based on relative sizes Reformulate classification process in terms ofReformulate classification process in terms of discriminant functions: x is assigned to Ck if x is assigned to Ck if

8/16/99 Bayes Decision Theory Discriminant function examples:

8/16/99 Bayes Decision Theory Discriminant function examples: 2-class problem

8/16/99 Bayes Decision Theory Why is such a big deal ?

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y  [/ah/, /eh/,.. /uh/] FFT melscale bank apple,...,zebra

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition FFT melscale bank /t/ /aal//aol//owl/

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition How do Humans do it?

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition “This machine can recognize speech” ??

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition “This machine can wreck a nice beach” !!

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y FFT melscale bank

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #1: Speech Recognition Why is such a big deal ? Example #1: Speech Recognition = x y FFT melscale bank P(“wreck a nice beach”) = P(“recognize speech”) = 0.02 Language Model

8/16/99 Bayes Decision Theory Why is such a big deal ? Example #2: Computer Vision Why is such a big deal ? Example #2: Computer Vision Low-Level Image Measurements High-Level Model Knowledge

8/16/99 Bayes Why is such a big deal ? Example #3: Curve Fitting Why is such a big deal ? Example #3: Curve Fitting E +  ln p(x|c) + ln p(c)

8/16/99 Bayes Why is such a big deal ? Example #4: Snake Tracking Why is such a big deal ? Example #4: Snake Tracking E +  ln p(x|c) + ln p(c)

8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction, Color Layers Statistical Models (Forsyth+Ponce Chap. 6, Duda+Hart+Stork: Chap. 1-5) - Bayesian Decision Theory - Density Estimation Computer Vision: Vision and Modeling

8/16/99 Probability Density Estimation Collect Data: x1,x2,x3,x4,x5,... x x ? Estimate:

8/16/99 Probability Density Estimation Parametric Representations Non-Parametric Representations Mixture Models

8/16/99 Probability Density Estimation Parametric Representations - Normal Distribution (Gaussian) - Maximum Likelihood - Bayesian Learning

8/16/99 Normal Distribution

8/16/99 Multivariate Normal Distribution

8/16/99 Multivariate Normal Distribution Why Gaussian ? Simple analytical properties: - linear transformations of Gaussians are Gaussian - marginal and conditional densities of Gaussians are Gaussian - any moment of Gaussian densities is an explicit function of  “Good” Model of Nature: - Central Limit Theorem: Mean of M random variables is distributed normally in the limit.

8/16/99 Multivariate Normal Distribution Discriminant functions:

8/16/99 Multivariate Normal Distribution Discriminant functions: equal priors + cov: Mahalanobis dist.

8/16/99 Multivariate Normal Distribution How to “learn” it from examples: Maximum Likelihood Bayesian Learning

8/16/99 Maximum Likelihood How to “learn” density from examples: x x ? ?

8/16/99 Maximum Likelihood Likelihood that density model   generated data X :

8/16/99 Maximum Likelihood Likelihood that density model   generated data X :

8/16/99 Maximum Likelihood Learning = optimizing (maximizing likelihood / minimizing E):

8/16/99 Maximum Likelihood Maximum Likelihood for Gaussian density: Close-form solution: