Lecture 26: Faces and probabilities

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CS479/679 Pattern Recognition Dr. George Bebis
Lecture 3 Nonparametric density estimation and classification
Applied Probability Lecture 5 Tina Kapur
Face Recognition and Detection
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Visual Recognition Tutorial
Object Detection and Recognition
How many object categories are there? Biederman 1987.
Object Recognition Szeliski Chapter 14.
Recognition Readings Szeliski Chapter 14 The “Margaret Thatcher Illusion”, by Peter Thompson.
Project 4 out today –help session today –photo session today Project 2 winners Announcements.
Machine Learning CMPT 726 Simon Fraser University
Visual Recognition Tutorial
Computer Vision I Instructor: Prof. Ko Nishino. Today How do we recognize objects in images?
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Project 2 due today Project 3 out today –help session today Announcements.
Exercise Session 10 – Image Categorization
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
So far focused on 3D modeling
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Jakob Verbeek December 11, 2009
ECE 471/571 – Lecture 2 Bayesian Decision Theory 08/25/15.
Nonparametric Density Estimation Riu Baring CIS 8526 Machine Learning Temple University Fall 2007 Christopher M. Bishop, Pattern Recognition and Machine.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Probability and Statistics in Vision. Probability Objects not all the sameObjects not all the same – Many possible shapes for people, cars, … – Skin has.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Intro. ANN & Fuzzy Systems Lecture 15. Pattern Classification (I): Statistical Formulation.
Recognition Readings – C. Bishop, “Neural Networks for Pattern Recognition”, Oxford University Press, 1998, Chapter 1. – Szeliski, Chapter (eigenfaces)
8/16/99 Computer Vision: Vision and Modeling. 8/16/99 Lucas-Kanade Extensions Support Maps / Layers: Robust Norm, Layered Motion, Background Subtraction,
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
CS Statistical Machine learning Lecture 7 Yuan (Alan) Qi Purdue CS Sept Acknowledgement: Sargur Srihari’s slides.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
CS479/679 Pattern Recognition Dr. George Bebis
Lecture 15. Pattern Classification (I): Statistical Formulation
Histograms CSE 6363 – Machine Learning Vassilis Athitsos
Announcements Project 2 due tomorrow night Project 3 out Wednesday
SFM under orthographic projection
Lecture 14: Introduction to Recognition
Special Topics In Scientific Computing
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
Lecture 25: Introduction to Recognition
Announcements Project 1 artifact winners
Outline Parameter estimation – continued Non-parametric methods.
LECTURE 05: THRESHOLD DECODING
Image Segmentation Techniques
LECTURE 05: THRESHOLD DECODING
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Parametric Estimation
CS4670: Intro to Computer Vision
Pattern Recognition and Machine Learning
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
CS4670: Intro to Computer Vision
Announcements Project 2 artifacts Project 3 due Thursday night
Announcements Project 4 out today Project 2 winners help session today
Parametric Methods Berlin Chen, 2005 References:
Where are we? We have covered: Project 1b was due today
Announcements Artifact due Thursday
Nonparametric density estimation and classification
LECTURE 05: THRESHOLD DECODING
Announcements Artifact due Thursday
Chapter 3: Maximum-Likelihood and Bayesian Parameter Estimation (part 2)
The “Margaret Thatcher Illusion”, by Peter Thompson
Bayesian Decision Theory
ECE – Pattern Recognition Lecture 4 – Parametric Estimation
Presentation transcript:

Lecture 26: Faces and probabilities CS4670/5670: Intro to Computer Vision Noah Snavely Lecture 26: Faces and probabilities

Face detection Do these images contain faces? Where?

Skin classification techniques Skin classifier Given X = (R,G,B): how to determine if it is skin or not? Nearest neighbor find labeled pixel closest to X choose the label for that pixel Data modeling fit a model (curve, surface, or volume) to each class Probabilistic data modeling fit a probability model to each class

Probability Basic probability X is a random variable P(X) is the probability that X achieves a certain value or Conditional probability: P(X | Y) probability of X given that we already know Y called a PDF probability distribution/density function a 2D PDF is a surface, 3D PDF is a volume continuous X discrete X

Probabilistic skin classification Now we can model uncertainty Each pixel has a probability of being skin or not skin Skin classifier Given X = (R,G,B): how to determine if it is skin or not? Choose interpretation of highest probability set X to be a skin pixel if and only if Where do we get and ?

Learning conditional PDF’s We can calculate P(R | skin) from a set of training images It is simply a histogram over the pixels in the training images each bin Ri contains the proportion of skin pixels with color Ri This doesn’t work as well in higher-dimensional spaces. Why not? Approach: fit parametric PDF functions common choice is rotated Gaussian center covariance orientation, size defined by eigenvecs, eigenvals

Learning conditional PDF’s We can calculate P(R | skin) from a set of training images It is simply a histogram over the pixels in the training images each bin Ri contains the proportion of skin pixels with color Ri But this isn’t quite what we want Why not? How to determine if a pixel is skin? We want P(skin | R), not P(R | skin) How can we get it?

Bayes rule In terms of our problem: The prior: P(skin) what we measure (likelihood) domain knowledge (prior) what we want (posterior) normalization term The prior: P(skin) Could use domain knowledge P(skin) may be larger if we know the image contains a person for a portrait, P(skin) may be higher for pixels in the center Could learn the prior from the training set. How? P(skin) could be the proportion of skin pixels in training set

Bayesian estimation likelihood posterior (unnormalized) Goal is to choose the label (skin or ~skin) that maximizes the posterior this is called Maximum A Posteriori (MAP) estimation = minimize probability of misclassification Suppose the prior is uniform: P(skin) = P(~skin) = 0.5 in this case , maximizing the posterior is equivalent to maximizing the likelihood if and only if this is called Maximum Likelihood (ML) estimation

Skin detection results

General classification This same procedure applies in more general circumstances More than two classes More than one dimension H. Schneiderman and T.Kanade Example: face detection Here, X is an image region dimension = # pixels each face can be thought of as a point in a high dimensional space H. Schneiderman, T. Kanade. "A Statistical Method for 3D Object Detection Applied to Faces and Cars". IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2000) http://www-2.cs.cmu.edu/afs/cs.cmu.edu/user/hws/www/CVPR00.pdf