Page 1 CVPR Workshop: Statistical Analysis in Computer Vision A Statistical Assessment of Subject Factors in the PCA Recognition of Human Subjects Geof.

Slides:



Advertisements
Similar presentations
Multiple Analysis of Variance – MANOVA
Advertisements

Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Page 1 The CSU Face Identification Evaluation System, ICVS 2003 Talk The CSU Face Identification Evaluation System: Its Purpose, Features and Structure.
1 Chapter 4 Experiments with Blocking Factors The Randomized Complete Block Design Nuisance factor: a design factor that probably has an effect.
Chapter 4 Randomized Blocks, Latin Squares, and Related Designs
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
Linear Regression t-Tests Cardiovascular fitness among skiers.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
Statistics 100 Lecture Set 7. Chapters 13 and 14 in this lecture set Please read these, you are responsible for all material Will be doing chapters
HSRP 734: Advanced Statistical Methods July 24, 2008.
Correlation and regression
Basic Data Analysis for Quantitative Research
Copyright © 2010, 2007, 2004 Pearson Education, Inc. *Chapter 29 Multiple Regression.
Class 17: Tuesday, Nov. 9 Another example of interpreting multiple regression coefficients Steps in multiple regression analysis and example analysis Omitted.
Regression Analysis. Unscheduled Maintenance Issue: l 36 flight squadrons l Each experiences unscheduled maintenance actions (UMAs) l UMAs costs $1000.
Lecture 14 Analysis of Variance Experimental Designs (Chapter 15.3)
Lecture 24: Thurs. Dec. 4 Extra sum of squares F-tests (10.3) R-squared statistic (10.4.1) Residual plots (11.2) Influential observations (11.3,
Lecture 24: Thurs., April 8th
Comparison and Combination of Ear and Face Images in Appearance-Based Biometrics IEEE Trans on PAMI, VOL. 25, NO.9, 2003 Kyong Chang, Kevin W. Bowyer,
Today Concepts underlying inferential statistics
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Week 9 Chapter 9 - Hypothesis Testing II: The Two-Sample Case.
Overview Definition Hypothesis
© 2013 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
Copyright © 2014, 2011 Pearson Education, Inc. 1 Chapter 25 Categorical Explanatory Variables.
Simple Covariation Focus is still on ‘Understanding the Variability” With Group Difference approaches, issue has been: Can group membership (based on ‘levels.
AP STATISTICS LESSON 11 – 2 (DAY 1) Comparing Two Means.
Chapter 9 Comparing More than Two Means. Review of Simulation-Based Tests  One proportion:  We created a null distribution by flipping a coin, rolling.
Covariance and correlation
Correlation.
Comparing Two Population Means
Statistics & Biology Shelly’s Super Happy Fun Times February 7, 2012 Will Herrick.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved OPIM 303-Lecture #9 Jose M. Cruz Assistant Professor.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide © 2012 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 1 Slide Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination n Model Assumptions n Testing.
Linear Functions 2 Sociology 5811 Lecture 18 Copyright © 2004 by Evan Schofer Do not copy or distribute without permission.
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
1 Chapter 3 Multiple Linear Regression Multiple Regression Models Suppose that the yield in pounds of conversion in a chemical process depends.
Lesson Significance Tests: The Basics. Vocabulary Hypothesis – a statement or claim regarding a characteristic of one or more populations Hypothesis.
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Chapter 13 Multiple Regression
Simple linear regression Tron Anders Moger
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Chapter 6: Analyzing and Interpreting Quantitative Data
Chapter 8: Simple Linear Regression Yang Zhenlin.
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 12 Multiple.
Business Statistics for Managerial Decision Farideh Dehkordi-Vakil.
Analyzing Expression Data: Clustering and Stats Chapter 16.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Business Research Methods
PART 2 SPSS (the Statistical Package for the Social Sciences)
Statistics for Managers Using Microsoft Excel, 5e © 2008 Prentice-Hall, Inc.Chap 14-1 Statistics for Managers Using Microsoft® Excel 5th Edition Chapter.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
Nonparametric Statistics
Biostatistics Regression and Correlation Methods Class #10 April 4, 2000.
Statistics 25 Paired Samples. Paired Data Data are paired when the observations are collected in pairs or the observations in one group are naturally.
STA248 week 121 Bootstrap Test for Pairs of Means of a Non-Normal Population – small samples Suppose X 1, …, X n are iid from some distribution independent.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Regression Analysis AGEC 784.
Two-way ANOVA with significant interactions
Correlation – Regression
Hypothesis Testing Review
COMPARING VARIABLES OF ORDINAL OR DICHOTOMOUS SCALES: SPEARMAN RANK- ORDER, POINT-BISERIAL, AND BISERIAL CORRELATIONS.
Presentation transcript:

Page 1 CVPR Workshop: Statistical Analysis in Computer Vision A Statistical Assessment of Subject Factors in the PCA Recognition of Human Subjects Geof Givens*, J. Ross Beveridge, Bruce A. Draper & David Bolme Computer Science, Colorado State University *Statistics, Colorado State University CVPR Workshop: Statistical Analysis in Computer Vision June 22, 2003

Page 2 CVPR Workshop: Statistical Analysis in Computer Vision Some Human Subjects Are Harder to Recognize than Others. Why?

Page 3 CVPR Workshop: Statistical Analysis in Computer Vision Better Formulate the Question Choose an algorithm –Something standard and simple. –PCA using our publicly released version. Choose a distance measure: –Yambor Angle is good and similar to Moon-Phillips FERET. –Much better than standards such as L2. Define & Collect Covariates: –Few covariates were collected initially with FERET images. –One person at CSU, David Bolme, scored all images. Big questions. –What to measure? –Covariates in isolation or together?

Page 4 CVPR Workshop: Statistical Analysis in Computer Vision Refinement of NIST preprocessing used in FERET. Some Basics: Image Preprocessing Integer to float conversion –Converts 256 gray levels to single- floats Geometric Normalization –Aligns human chosen eye coordinates Masking –Crop with elliptical mask leaving only face visible. Histogram Equalization –Histogram equalizes unmasked pixels: 256 levels. Pixel normalization –Shift and scale pixel values so mean pixel value is zero and standard deviation over all pixels is one.

Page 5 CVPR Workshop: Statistical Analysis in Computer Vision More Basics: Standard PCA Algorithm. … PCA space projection Training images Eigenspace Training Testing … Distance Matrix Remove in/out training issue by training on all images.

Page 6 CVPR Workshop: Statistical Analysis in Computer Vision What Covariates? Glasses Bangs Facial Hair Mouth Smiling? Eyes AgeGenderRace

Page 7 CVPR Workshop: Statistical Analysis in Computer Vision The Final Set of Subject Covariates

Page 8 CVPR Workshop: Statistical Analysis in Computer Vision Collecting the Covariates

Page 9 CVPR Workshop: Statistical Analysis in Computer Vision What to Measure? Recognition Rate on Partitioned Data Measure Recognition Rate for Partitioned Images. –Partition images by covariate : e.g. male versus female. –Compare recognition rate on different sets. Good –Answers a very specific version of the question. –Recognition rate is a standard performance measure. Bad –Fails to adjust, i.e. control for, other covariates. –Recognition rate for a probe set, not a single subject. –Hidden dependence on gallery image set. –Statistical significance is hard to interpret.

Page 10 CVPR Workshop: Statistical Analysis in Computer Vision What to measure? Intrapersonal Image Pair Distance Measure Distance (Similarity) Between Images – For two images of one subject, closer is better. Good –Measure is independent of other subjects, probe sets, etc. –Measure is continuous, analysis of variance appropriate. –Linear model accounts for all covariates at one time. –Standard tests of statistical significance apply. Bad –Distance to recognition rate connection indirect. –Supplemental analysis required to establish linkage. Made through intermediate measure: recognition rank.

Page 11 CVPR Workshop: Statistical Analysis in Computer Vision A Glimpse of Distance and Rank Data

Page 12 CVPR Workshop: Statistical Analysis in Computer Vision Image Pairs for Three Example Subjects Shown Above Recognition Rank = 1 Distance = -299*Distance = -206*Distance = -110* Distance x 100,000 *

Page 13 CVPR Workshop: Statistical Analysis in Computer Vision Linear Model Relating Distance to Subject Covariates Y i = Distance metric for image pair i. X i = Human covariate factors for image pair i.  = Parameters quantifying factor effects. Y i =  0 +  1 X i1 +  2 X i2 + … +  i with  i ~ iid Normal(0,  2 )

Page 14 CVPR Workshop: Statistical Analysis in Computer Vision Comments on ANOVA 1,072 Subject image pairs. –Each image pair corresponds to a unique person. Both images from the same day –Delay between images is known to make recognition harder. Subjects did not add/remove glasses –Pilot studies included this case, it is harder. ANOVA yielded R 2 = 0.39 –Covariates explain 39% of observed variation. –Notable given 75% of subjects can be recognized at rank 1.

Page 15 CVPR Workshop: Statistical Analysis in Computer Vision FERET Subject Covariates Summary of Results Glasses Off Expression Changes Makeup Changes Mouth Changes Glasses Always On Age YoungAge Old Eyes OpenEyes Open/ClosedEyes Always Closed Expression Neutral Always Non-neutral Race White Race Asian Race African-Amer. Race Other No Facial Hair Always Facial HairFacial Hair Changes MaleFemale -50%-40%-30%-20%-10%0% 10%20%30%40%50% Change in Similarity Measure Harder to Recognize Easier to Recognize No MakeupAlways Makeup Mouth Closed Mouth Always Open No Bangs Always Bangs Bangs Change Skin ClearSkin Not Clear Base-Case

Page 16 CVPR Workshop: Statistical Analysis in Computer Vision Two Possible Reservations Unbalanced training biases conclusions?Unbalanced training biases conclusions? “Asians are closer to each other than whites because the algorithm was trained on very few of them. The unbalanced training means less well represented sets of samples are not as well separated.” Choice of Y = distance renders results irrelevant?Choice of Y = distance renders results irrelevant? “We care about recognition rank, or rank-k recognition rate. I don’t know whether your response variable is strongly related to this. Even if it is related overall, does the relationship hold for specific groups?”

Page 17 CVPR Workshop: Statistical Analysis in Computer Vision Plot of Distance versus the Log of Recognition Rank Distance (x 100,000) Log of Recognition Rank 845 out of 1,072 subject image pairs are closest matches: i.e. Rank is 1. Visual inspection suggests strong relationship between distance and rank.

Page 18 CVPR Workshop: Statistical Analysis in Computer Vision Modeling the Relationship Between Distance and Recognition Rank Y i = Was the ith image pair matched at rank 1 ? (i.e. Y i = 1 if R i = 0 and otherwise Y i = 0) X i = Distance metric for image pair i.  = Parameters quantifying the relationship. g(  Yi|Xi ) = X i '  =  0 +  1 X i Y i | X i ~ f(  Yi|Xi ) independently Now: g(z) = log (z/(1-z)), f(  Yi|Xi ) = Bernoulli(  Yi|Xi )

Page 19 CVPR Workshop: Statistical Analysis in Computer Vision Logistic Regression of Rank Indicator on Distance Distance (x 100,000) Z i - One if Rank is One The probability of a rank 1 match decreases sharply with increasing distance;  1 is with p-value = Similar when grouped for race, gender, age, skin and glasses.

Page 20 CVPR Workshop: Statistical Analysis in Computer Vision Supplemental Balanced Training Experiment for Race. YoungOld Asian896 White896 Black7816 White7816 Other626 White626 To Test Race p-value Asian38078 Glasses = Off, Eyes = Open, Skin = Clear Balance Age √ √ √ Black Other TestCompareConfirm PCA Dim.Total Images Experiment 720 White 143 Asian 121 Black 88 Other Follow-up experiment replicates the first, but with carefully balanced training. Distribution of race in 1,072 Subjects is not balanced.

Page 21 CVPR Workshop: Statistical Analysis in Computer Vision Supplemental Balanced Training Experiments for Other Covariates CompareClearOther Old13157 Young13157 OldYoung Clear7257 Other7257 OldYoung Off1416 On1416 To Test Age 50 Glasses = Off, Eyes = Open, Race = White To Test Skin Glasses = Off, Eyes = Open, Race = White To Test Glasses √ √ Compare 5 6 Balance Skin Balance Age Eyes = Open, Race = White, Skin = Clear Age Skin Glasses Test √ TestCompare Total Images PCA Dim.Total Images 117 PCA Dim. Total Images Confirmp-value Confirmp-value Confirmp-value Experiment

Page 22 CVPR Workshop: Statistical Analysis in Computer Vision Sampling of Related Work Nicholas Furl, P. Jonathon Phillips and Alice J. O’Toole. Face recognition algorithms and the other-race effect: computational mechanisms for a developmental contact hypothesis. PCA-based study, recall style experiment, looked at White vs. Asian distinction in FERET, training bias like that of humans. Jeffrey F. Cohn Ralph Gross, Jianbo Shi. Quo vadis face recognition?: The current state of the art in face recognition, 2002 Partitioned probe set of 1,119 subjects: 87.6 vs recognition rate for females vs. males. Used FaceIt and not FERET data. Found men easier. Face Recognition Vendor Test 2002, Evaluation Report, P. Jonathon Phillips, Patrick Grother, Ross J. Micheals, Duane M. Blackburn, Elham Tabassi, Mike Bone Large Data Set, Eight Systems, CMC and ROC analysis. Gender test partitioned probe set and found men easier.

Page 23 CVPR Workshop: Statistical Analysis in Computer Vision More on the Gender Effect, or Lack of Gender Effect. From FRVT 2002 Evaluation Report These results are based upon simple partition of probe set.

Page 24 CVPR Workshop: Statistical Analysis in Computer Vision What Happens if We Replicate Simpler Partition on Gender? Fit our data with a simple on-way ANOVA LM on Gender. Analogous to partition of male vs. female probes. Result –“statistically significant” gender effect. –Male subjects about 13% more easier: p < But –We already know this result is wrong! –Full analysis with all our covariates shows no gender effect. This exercise illustrates a classic mistake It is far too much to hope that a ‘sample of convenience’ will be balanced with respect to every confounding variable. What might be confounding?

Page 25 CVPR Workshop: Statistical Analysis in Computer Vision Conclusions Most comprehensive study of its kind to date. –i.e. most covariates considered together. Interesting Discoveries –Glasses aid recognition (but don’t take them off). –Non-white subjects easier, despite smaller sample. –Lack of gender effect, why? Was there enough data? –Yes, compares favorably to standards for statistical analysis. Hubris –All studies have limitations. –Linkage between distance and recognition performance. Future work –Analogous experiment on portion of FRVT 2002 data. –Image covariate analysis.

Page 26 CVPR Workshop: Statistical Analysis in Computer Vision A Little Extra: The Best and Worst of FERET Best Subjects (Closest Pairs) Worst Subjects (Most Distant Pairs)