School of Electrical and Computer Engineering A Mathematical Theory of Automatic Target Recognition Aaron D. Lanterman

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Learning deformable models Yali Amit, University of Chicago Alain Trouvé, CMLA Cachan.
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
QR Code Recognition Based On Image Processing
Hypothesis testing Another judgment method of sampling data.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CS479/679 Pattern Recognition Dr. George Bebis
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Aaron D. Lanterman School of Electrical and Computer Engineering Target Recognition in Cluttered Infrared Scenes via.
NASSP Masters 5003F - Computational Astronomy Lecture 5: source detection. Test the null hypothesis (NH). –The NH says: let’s suppose there is no.
Introduction To Tracking
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
What is Statistical Modeling
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Visual Recognition Tutorial
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Maximum likelihood (ML) and likelihood ratio (LR) test
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
1 On the Statistical Analysis of Dirty Pictures Julian Besag.
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Highlights Lecture on the image part (10) Automatic Perception 16
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Maximum likelihood (ML)
Noise Estimation from a Single Image Ce Liu William T. FreemanRichard Szeliski Sing Bing Kang.
Crash Course on Machine Learning
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
1 Patch Complexity, Finite Pixel Correlations and Optimal Denoising Anat Levin, Boaz Nadler, Fredo Durand and Bill Freeman Weizmann Institute, MIT CSAIL.
An algorithm for dynamic spectrum allocation in shadowing environment and with communication constraints Konstantinos Koufos Helsinki University of Technology.
Bala Lakshminarayanan AUTOMATIC TARGET RECOGNITION April 1, 2004.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.
OBJECT RECOGNITION. The next step in Robot Vision is the Object Recognition. This problem is accomplished using the extracted feature information. The.
Topic 10 - Image Analysis DIGITAL IMAGE PROCESSING Course 3624 Department of Physics and Astronomy Professor Bob Warwick.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Chapter 21 R(x) Algorithm a) Anomaly Detection b) Matched Filter.
Automated Detection and Classification Models SAR Automatic Target Recognition Proposal J.Bell, Y. Petillot.
03/05/03© 2003 University of Wisconsin Last Time Tone Reproduction If you don’t use perceptual info, some people call it contrast reduction.
Forward-Scan Sonar Tomographic Reconstruction PHD Filter Multiple Target Tracking Bayesian Multiple Target Tracking in Forward Scan Sonar.
Lecture notes for Stat 231: Pattern Recognition and Machine Learning 3. Bayes Decision Theory: Part II. Prof. A.L. Yuille Stat 231. Fall 2004.
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
First topic: clustering and pattern recognition Marc Sobel.
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Expectation-Maximization (EM) Case Studies
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Automated Detection and Classification Models SAR Automatic Target Recognition Proposal J.Bell, Y. Petillot.
BCS547 Neural Decoding.
12/4/981 Automatic Target Recognition with Support Vector Machines Qun Zhao, Jose Principe Computational Neuro-Engineering Laboratory Department of Electrical.
SAR-ATR-MSTAR TARGET RECOGNITION FOR MULTI-ASPECT SAR IMAGES WITH FUSION STRATEGIES ASWIN KUMAR GUTTA.
Looking at people and Image-based Localisation Roberto Cipolla Department of Engineering Research team
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Automated Detection and Classification Models SAR Automatic Target Recognition Proposal J.Bell, Y. Petillot.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
MultiModality Registration Using Hilbert-Schmidt Estimators By: Srinivas Peddi Computer Integrated Surgery II April 6 th, 2001.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Lecture 1.31 Criteria for optimal reception of radio signals.
- photometric aspects of image formation gray level images
Bag-of-Visual-Words Based Feature Extraction
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Solving an estimation problem
Bayesian inference J. Daunizeau
Presentation transcript:

School of Electrical and Computer Engineering A Mathematical Theory of Automatic Target Recognition Aaron D. Lanterman

What Makes ATR “Harder” than Factoring Large Numbers? Factoring large numbers may be NP- hard, but... At least it’s easy to precisely specify what the problem is! Not so easy in ATR –Subject to controversy

Can You Build an Airplane Without a Theory of Aerodynamics? Sure! Without aerodynamic theory, you can do this... …but with a theory, you can do this!

Can You Build an Communication Systems w/out Information Theory? Sure! Without Information Theory, you can do this… …but with Information Theory, you can do this!

Dick Blahut likens the situation to steam engines coming before the science of thermodynamics First steam engines build by entrepreneurs and “inventors” –Thomas Savery: 17 th and 18 th centuries –Necessity the mother of invention! Thermodynamics didn’t begin to crystallize until mid 19 th century… but with it, you eventually get Steam Engines and Thermodynamics

Before Shannon, your boss might ask you to do the impossible, and fire you if you failed to do it! Your boss cannot fire your for failing to exceed channel capacity! You can tell your boss you need a better channel 1948: Claude Shannon’s “A Mathematical Theory of Communication” (1948) –Later renamed “The Mathematical Theory of Communication” Found fundamental limits on what is possible, i.e. channel capacity Shannon’s Lightning Bolt shouldn’t

Theory and Technology Advances in theory are not enough; also need the technology –Aerodynamic theory alone won’t get you a B-2; need advances in materials, manufacturing –Information theory along won’t get you cell phones; need fast DSP chips, good batteries, even more theory (i.e. coding theory) Theory tells you what’s possible, but sometimes only hints at how to get there –Quantum computing folks: does this sound familiar?

Info-Theoretic View of ATR Source ChannelDecoder Hypothesis testing (LRT, GLRT) ML, Bayes, Neyman Pearson Estimation ML, MAP, M.M.S.E., Bayes Miss, false alarm rate Confusion matrices Bias, Variance, M.S.E. Optimality Criteria Performance Bounds Chernoff Stein’s Lemma Cramer-Rao Target Recognizer Scene Understanding Multiple Sensors Scene Synthesizer Database (Statistical Estimation-Theoretic) CIS/MIM

What Makes ATR “Harder” than Designing a Cell Phone? The space of X for real-world scenes is extremely complicated You don’t get to pick p(x) Likelihood p(y|x) is difficult to formulate –The “channel” is often deliberately hostile Targets hiding in clutter Using decoys and camouflage Radars can be subject to jamming

Geometric variability –Position –Orientation –Articulation –“Fingerprint” Environmental variability –Thermal variability in infrared –Illumination variability in visual Complexity variability –Number of objects not known Variability in Complex Scenes

Ulf Grenander Student of Cramér (yes, that Cramér) PhD on statistical inference in function spaces (1950) “Toeplitz Forms and their Applications” (with Szegö) –Fundamental work on spectral estimation (1958) “Probabilities on Algebraic Structures” (1968) “Tutorial on Pattern Theory” - unpublished manuscript –Inspired classic paper by Geman & Geman (1983)

General Pattern Theory Generalize standard probability, statistics, and shape theory Put probability measures on complex structures –Biological structures Mitochondria Amoebas Brains Hippocampus –Natural language –Real-world scenes of interest in ATR

The 90’s GPT Renaissance Made possible by increases in computer power Michael Miller (Washington Univ., now at JHU) did a sabbatical with Grenander Fields Medalist David Mumford moves from Harvard to Brown; shifts from algebraic geometry to pattern theory

Composite Parameter Spaces Move away from thinking of detection, location, recognition, etc. as separate problems Naturally handles obscuration Don’t know how many targets are in the scene in advance

Applying the Grenander Program (1) Take a Bayesian approach Many ATR algorithms seek features that are invariant to pose (position and orientation) Grenander’s Pattern Theory treats pose as nuisance variable in the ATR problem, and deals with it head on –Co-estimate pose, or integrate it out –At a given viewing angle, Target A at one orientation may look much like Target B at a different orientation –“…the nuisance parameter of orientation estimation plays a fundamental role in determining the bound on recognition” - Grenander, Miller, & Srivastava U. Grenander, M.I. Miller, and A. Srivastava, “Hilbert-Schmidt Lower Bounds for Estimators on Matrix Lie Groups for ATR,” IEEE Trans. PAMI, Vol. 20, No. 2, Aug. 1998, pp

Applying the Grenander Program (2) Develop statistical likelihood Data fusion is natural At first, use as much of the data as possible –Be wary of preprocessing: edge extraction, segmentation etc. –Processing can never add information Data processing inequality from information theory If you need to extract features, i.e. for real-time computational tractability, try to avoid as much loss of information as possible

Analytic Performance Bounds Estimation bounds on continuous parameters –Cramér-Rao bounds for continuous pose parameters –Hilbert-Schmidt metrics for orientation parameters Bounds on detection/recognition probabilities –Stein’s Lemma, Chernoff bounds –Asymptotic analysis to approximate probabilities of error –Performance in a binary test is dominated by a term exponential in a distance measure between a “true” and an “alternate” target Adjust pose of “alternate” target to get closest match to “true” target as seen by the sensor system –Secondary term involving CRB on nuisance parameters Links pose estimation and recognition performance U. Grenander, A. Srivastava, and M.I. Miller, “Asymptotic Performance Analysis of Bayesian Target Recognition,” IEEE Trans. Info. Theory, Vol. 46, No. 4, July 2000, pp Anuj Srivastava

Reading One of DARPA’s BAAs… DARPA’s E3D program seeks: –“efficient techniques for rapidly exploiting 3-D sensor data to precisely locate and recognize targets.” BAA full of demands (hopes?) for different stages of the program, such as: –“The Target Acquisition and Recognition technology areas will develop techniques to locate and recognize articulating, reconfigurable targets under partial obscuration conditions, with an identification probability of 0.85%, a target rejection rate less than 5%, and a processing time of 3 minutes per target or less”

…Leads Us to Wondering If such a milestone is not reached, is that the fault of the algorithm or the sensor? –How does the DARPA Program Manager know who to fire? –Without a theory, the DARPA PM may fire someone who was asked to “exceed channel capacity,” i.e. given an impossible task What performance from a particular sensor is necessary to achieve a certain level of ATR performance, independent of the question of what algorithm is used?

Perspective Projection

Optical PSF Poisson Photocounting Noise Dead and Saturated Pixels Sensor Effects

Loglikelihood where Cascade with Sensor fusion natural; just add loglikelihoods CCD loglikelihood of Snyder et. al

Langevin Diffusion Processes Fix number of targets and target types Simulate Langevin diffusion: Distribution of Computed desired statistics from the samples Generalizes to non-Euclidean groups like rotations Gradient computation –Numeric approximations –Easy and fast on modern 3-D graphics hardware Write posterior in Gibbs form:

Birth Death Type-change Jump Processes

Gibbs style –Sample from a restricted part of the posterior Metropolis-Hastings style –Draw a “proposal” from a “proposal density” –Accept (or reject) the proposal with a certain probability Jump Strategies

Example Jump-Diffusion Process

Average Static State Average Dynamic State Thermal Variability Simulations from PRISM: Discretizes target surface using regions from CAD template and internal heat transfer model CIS/MIM

Can’t Hide from Thermal Variations Profile 8 Profile 45 Profile 75 Profile 140 Performance Variations Due To Thermodynamic Variability Performance Loss Due To Inaccurate Thermodynamic Information Cooper, Miller SPIE 97 CIS/MIM

Principle Component Representation of Thermal State Model radiance as scalar random field on surface Compute empirical mean & covariance from database of 2000 radiance profiles Karhunen-Loeve expansion using eigenfunctions of covariance on surface - “Eigentanks” Add expansion coefficients to parameter space –Fortunately, able to estimate directly given pose SPIE 97 Cooper, Grenander, Miller, Srivastava A younger, much thinner Aaron Lanterman Matt Cooper (now with Xerox) CIS/MIM

Meteorological VariationOperational Variation Composite Mode of Variation SPIE 97 Cooper, Grenander, Miller, Srivastava The First “Eigentanks” Remember, we’re showing 2-D views of full 3-D surfaces CIS/MIM

Joint MAP Est. of Pose and Thermal Signature Real NVESD M60 data (courtesy James Ratches) SPIE 98 Cooper and Miller Initial Estimate Final Estimate CIS/MIM

“Cost” of Estimating Thermal State MSE Performance Loss Comanche SNR = 5.08 dB CIS/MIM

Ladar/IR Sensor Fusion MSE Performance BoundInformation Bound LADAR (range) FLIR (intensity) Joe Kostakis Tom Green Jeff Shapiro CIS/MIM

LADAR & IR Sensor Fusion LADAR/FLIR Hannon Curve 15 degrees error LADAR/FLIR Hannon Curve 9 degrees error SPIE 98 Advanced Techniques ATR III Kostakis, Cooper, Green, Miller, OSullivan, Shapiro Snyder CIS/MIM

Target Models Panzer II Light Tank Hull Length: 4.81 m Width: 2.28 m Height: 2.15 m Sturmgeschultz III Self-Propelled Gun Hull Length: 6.77 m Width: 2.95 m Height: 2.16 m Semovente M41 Self-Propelled Gun Hull Length: m Width: 2.2 m Height: 2.15 m M48 A3 Main Battle Tank Hull Length: m Width: 3.63 m Height: m (Info and Top Row of Images from 3-D Ladar Challenge Problem Slides by Jacobs Sverdrup)

CR-Bound on Orientation Strum Semo Position assumed known Position unknown, must be co-estimated Interesting knee at 0.2 meters We take a performance hit!

M48 vs. Others M48 and Panzer have dissimilar signatures; most easily distinguished M48 and Semo have similar signatures; most easily confused

Semovente vs. Others At higher resolutions, Semo and M48 have most dissimilar signatures; most easily distinguished (perhaps there are nice features which only become apparent at higher resolutions?) Semo and Sturm have similar signatures; most easily confused At lower resolutions, Semo and Panzer have most dissimilar signatures; most easily distinguished

Joseph O’Sullivan Synthetic Aperture Radar Michael DeVore MSTAR Data Set Conditionally Gaussian model for pixel values with variances trained from data Likelihood based classification Target orientation unknown and uniformly distributed over 360 ° of azimuth Joint orientation estimation and target classification Train on 17° depression angle Test on 15° depression angle SAR Images Variance Images T72 BMP 2 CIS/MIM

Results using 72 variance images per target of 10° each, and using 80 x 80 pixel sub- images to reduce background clutter Probability of correct classification: 98% Average orientation error: < 10° Supported by ARO Center for Imaging Science DAAH and ONR MURI N Orientation MSE effects ID! CIS/MIM

Caveat Do not confuse the model with reality.

Where Should Clutter Go? (1) A “forward model,” i.e. a “scene simulator” A forest might go well in the “noise” part… non-Gaussian minimax entropy texture models by Song Chun Zhu

Where Should Clutter Go? (2) …but downtown Baghdad will not “whiten” Structured clutter is the most vexing May need to go in here, and directly manipulate the clutter …or a bit of each Where to draw the line?

Acknowledgments Much of the work described here was funded by the ARO Center for Imaging Science Also ONR (William Miceli) and AFOSR (Jon Sjogren) Slides with CIS/MIM tag were adapted from slides provided by Michael Miller