Frequentist and Bayesian Measures of Association Quality in Algorithmic Toolmark Identification.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Review bootstrap and permutation
Evaluating Classifiers
Pattern Recognition and Machine Learning
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
What they are and how they fit into
Current Research in Forensic Toolmark Analysis Helping to satisfy the “new” needs of forensic scientists with state of the art microscopy, computation.
Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
CCI Firearms and Toolmark Examiner Academy Workshop on Current Firearms and Toolmark Research Pushing Out the Frontiers of Forensic Science.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
MCS Multiple Classifier Systems, Cagliari 9-11 June Giorgio Valentini Random aggregated and bagged ensembles.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Deciding, Estimating, Computing, Checking How are Bayesian posteriors used, computed and validated?
1 How to be a Bayesian without believing Yoav Freund Joint work with Rob Schapire and Yishay Mansour.
Machine Learning CMPT 726 Simon Fraser University
End of Chapter 8 Neil Weisenfeld March 28, 2005.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 7 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Sequence comparison: Significance of similarity scores Genome 559: Introduction to Statistical and Computational Genomics Prof. James H. Thomas.
Ensemble Learning (2), Tree and Forest
Bootstrap spatobotp ttaoospbr Hesterberger & Moore, chapter 16 1.
Interval Estimates A point estimate gives a plausible single number estimate for a parameter. We may also be interested in a range of plausible values.
Crash Course on Machine Learning
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
Modeling Menstrual Cycle Length in Pre- and Peri-Menopausal Women Michael Elliott Xiaobi Huang Sioban Harlow University of Michigan School of Public Health.
ROOT: A Data Mining Tool from CERN Arun Tripathi and Ravi Kumar 2008 CAS Ratemaking Seminar on Ratemaking 17 March 2008 Cambridge, Massachusetts.
Using car4ams, the Bayesian AMS data-analysis code V. Palonen, P. Tikkanen, and J. Keinonen Department of Physics, Division of Materials Physics.
Current Research in Forensic Toolmark Analysis Petraco Group.
Computational Strategies for Toolmarks:. Outline Introduction Details of Our Approach Data acquisition Methods of statistical discrimination Error rate.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
Statistics In HEP Helge VossHadron Collider Physics Summer School June 8-17, 2011― Statistics in HEP 1 How do we understand/interpret our measurements.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
Introduction to Inferece BPS chapter 14 © 2010 W.H. Freeman and Company.
Univariate Linear Regression Problem Model: Y=  0 +  1 X+  Test: H 0 : β 1 =0. Alternative: H 1 : β 1 >0. The distribution of Y is normal under both.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Discriminant functions are trained on a finite set of data How much fitting should we do? What should the model’s dimension be? Model must be used to.
Academic Research Academic Research Dr Kishor Bhanushali M
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
- 1 - Overall procedure of validation Calibration Validation Figure 12.4 Validation, calibration, and prediction (Oberkampf and Barone, 2004 ). Model accuracy.
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
G. Cowan Lectures on Statistical Data Analysis Lecture 8 page 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem 2Random variables and.
Applied Quantitative Analysis and Practices LECTURE#14 By Dr. Osman Sadiq Paracha.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Anders Nielsen Technical University of Denmark, DTU-Aqua Mark Maunder Inter-American Tropical Tuna Commission An Introduction.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
We are interested in methods that produce an interval: Common interval methods for: Confidence intervals Prediction intervals Tolerance intervals Credibility/Probability.
Forensic Surface Metrology
Pushing Out the Frontiers of Forensic Science
Approaches to Physical Evidence Research In the Forensic Sciences
Machine Learning – Classification David Fenyő
(5) Notes on the Least Squares Estimate
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Special Topics In Scientific Computing
Overview of Supervised Learning
Statistical Process Control
Statistics Presentation
Quantifying uncertainty using the bootstrap
Discrete Event Simulation - 4
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Let’s do a Bayesian analysis
CS639: Data Management for Data Science
Introductory Statistics
Machine Learning in Business John C. Hull
Presentation transcript:

Frequentist and Bayesian Measures of Association Quality in Algorithmic Toolmark Identification

Outline Introduction Details of Our Approach The Data Some alternative (testable!) measures of an association quality Confidence: Vovk et al. Conformal Prediction Believability: Efron 2-groups “Empirical Bayes” Future directions

All impressions made by tools and firearms can be represented as numerical patterns – Machine learning trains a computer to recognize patterns Can give “…the quantitative difference between an identification and non-identification” Moran Can yield identification error rate estimates May be even confidence measures for I.D.s…… Background Information

Obtain striation/impression patterns from 3D microscopy Store files in ever expanding database (NIST Zheng ) Data files are available to practitioner and researcher community (NIST Zheng ) Data Acquisition

9mm-Glock Fired Cartridges

2D profiles 3D surfaces (interactive) Screwdriver Striation Patterns in Lead

We can simulate profiles as well Baiker Profile Simulator Based on DWT MRA May shed light on the processes which generate the surfaces

Working on a simulator for 2D toolmarks: Surface Simulator LH4 HL4 HH4 + Simulate stocastic detail

Toolmarks (screwdriver striation profiles) form database Biasotti-Murdock Dictionary For Striated Toolmarks: Consecutive Matching Striae Space CMS-space features

Visually explore: 3D PCA of CMS-space for 1740 real and simulated screwdriver striation patterns : ~9% variance retained

How good of a “match” is it? Conformal Prediction Vovk Data should be IID but that’s it Cumulative # of Errors Sequence of Unk Obs Vects 80% confidence 20% error Slope = % confidence 5% error Slope = % confidence 1% error Slope = 0.01 Can give a judge or jury an easy to understand measure of reliability of classification result This is an orthodox “frequentist” approach Roots in Algorithmic Information Theory Confidence on a scale of 0%-100% Testable claim: Long run I.D. error- rate should be the chosen significance level

How Conformal Prediction works for us Given a “bag” of obs with known identities and one obs of unknown identity Vovk Estimate how “wrong” labelings are for each observation with a non- conformity score (“wrong-iness”) Looking at the “wrong-iness” of known observations in the bag: Does labeling-i for the unknown have an unusual amount of “wrong-iness”??: For us, one-vs-one SVMs: If not: p possible-ID i ≥ chosen level of significance Put ID i in the (1 - )*100% confidence interval

Conformal Prediction Theoretical (Long Run) Error Rate: 5% Empirical Error Rate: 5.3% 14D PCA-SVM Decision Model for screwdriver striation patterns For 95%-CPT (PCA-SVM) confidence intervals will not contain the correct I.D. 5% of the time in the long run Straight-forward validation/explanation picture for court

An I.D. is output for each questioned toolmark This is a computer “match” What’s the probability the tool is truly the source of the toolmark? Similar problem in genomics for detecting disease from microarray data They use data and Bayes’ theorem to get an estimate How good of a “match” is it? Efron Empirical Bayes’

Bayesian Statistics The basic Bayesian philosophy: Prior Knowledge × Data = Updated Knowledge A better understanding of the world Prior × Data = Posterior

Empirical Bayes’ From Bayes’ Theorem we can get Efron : Estimated probability of not a true “match” given the algorithms' output z-score associated with its “match” Names: Posterior error probability (PEP) Kall Local false discovery rate (lfdr) Efron Suggested interpretation for casework: = Estimated “believability” that the specific tool produced the toolmark

Empirical Bayes’ Bootstrap procedure to get estimate of the KNM distribution of “Platt-scores” Platt,e1071 Use a “Training” set Use this to get p-values/z-values on a “Validation” set Inspired by Storey and Tibshirani’s Null estimation method Storey z-score From fit histogram by Efron’s method get: “mixture” density We can test the fits to and ! What’s the point?? z-density given KNM => Should be Gaussian Estimate of prior for KNM Use SVM to get KM and KNM “Platt-score” distributions Use a “Validation” set

Rough Procedure Sample to get a set of IID simulated log(KNM-scores) (“reusing data” less too…??) Compute p-values for the validation set from the fit null Compute KM scores off the validation set

Use locfdr locfdr Fit classic Poisson regression for f(z) Use modified locfdr/JAGS JAGS,Plummer or Stan Stan Fit Bayesian hierarchal Poisson regressions z z Fit local-fdr models Check: Is underneath here about N(0,1)?

Check Calibration The SVM alg got these Primer shear IDs wrong Do right answers get high “believability” (low pep?) Do wrong answers get low “believability” (high pep?)

Posterior Association Probability: Believability Curve 12D PCA-SVM locfdr fit for Glock primer shear patterns +/- 2 standard errors

Empirical Bayes’ Model’s use with crime scene “unknowns”: This is the est. post. prob. of no association = = 0.027% Computer outputs “match” for: unknown crime scene toolmarks-with knowns from “Bob the burglar” tools This is an uncertainty in the estimate

Future Directions Exploit scientific image processing toolkits openCV, scikit-image, vlfeat (features) openCV, Panorama Tools, Hugin, Montage Image Mosaic (stitching) Better toolmark features CADRE Research Labs doing impressive work Parallel implementation of computationally intensive routines Forensic algorithm architectural review board (ARB) for law applications akin to Khronos group Khronos or Boost.org Boost OpenFMC

Acknowledgements Alan Zheng (NIST) Erich Smith (If I tell you where he works, I have to kill you) Ryan Lillian (CADRE) JoAnn Buscaglia (You all know where she works) Professor Chris Saunders (SDSU) Research Team: Ms. Tatiana Batson Dr. Martin Baiker Ms. Julie Cohen Dr. Peter Diaczuk Mr. Antonio Del Valle Ms. Carol Gambino Dr. James Hamby Mr. Nick Natalie Mr. Mike Neel Ms. Alison Hartwell, Esq. Ms. Loretta Kuo Ms. Frani Kammerman Dr. Brooke Kammrath Mr. Chris Lucky Off. Patrick McLaughlin Dr. Linton Mohammed Ms. Diana Paredes Mr. Nicholas Petraco Ms. Stephanie Pollut Dr. Peter Pizzola Dr. Jacqueline Speir Dr. Peter Shenkin Mr. Chris Singh Mr. Peter Tytell Dr. Peter Zoon

Website: Data, codes, reprints and preprints: