Download presentation
Presentation is loading. Please wait.
Published byLionel Pitts Modified over 9 years ago
1
Kevin Cherry Robert Firth Manohar Karki
2
Accurate detection of moving objects within scenes with dynamic background, in scenarios where the camera is mostly stationary. Problem Definition
3
Prerequisite for object tracking and recognition Motivation
4
Methods that employ local (pixel-wise) models of intensity Methods that have regional models of intensity Previous Work– modelling intensity
5
Previous Work – Background Subtraction Naïve approach: | frame i – background | > threshold Better:| frame i – μ | > kσ
6
Background – Bayes' theorem APPENDIX: EXAMPLEEXAMPLE
7
Background – Minimum Cut
8
Background – Probability Density Estimator (parametric)
9
Background – Kernel Density Estimator (nonparametric)
10
Mean Squared Error: Balloon Estimator: Sample-point Estimator: For computation reduction, OR, APPENDIX – Bandwidth Matrix Classes Background - Bandwidth Estimation
11
Background: Markov Random Field
12
Guy flying kite. 15 consecutive frames merged using Photoshop. Background– Temporal Persistence
13
Method Overview
14
Domain: (x, y) coordinates Range: (r, g, b) color values at each (x, y) coordinate Joint Domain-Range Representation: f R,G,B,X,Y (r, g, b, x, y) Directly models dependencies between neighboring pixels Joint Domain-Range Representation Examples:
15
Modeling the Background Background pixels: Y[0]= [15, 15, 15, 1, 1]; Y[1]= [10,12,12.8,1,1]; Y[2]= [ 16,1,2,1,1]; Y[3]= [16, 10, 13, 1, 1]; Bandwidth matrix H: 16 0 0 0 0 0 16 0 0 0 0 0 16 0 0 0 0 0 31 0 0 0 0 0 21 Inverse of H: 0.063 0 0 0 0 0 0.063 0 0 0 0 0 0.063 0 0 0 0 0 0.032 0 0 0 0 0 0.048 |H| = 2666495 |H| -1/2 = 0.00061239 x = [16, 14, 15, 2, 1]; d = x – y[0] = [16-15,14-15,15-15,2-1,1-1] = [1,-1,0,1,0] d T H -1 d= 0.1573 Φ H = 0.00061239 * (2π) -5/2 * exp(-1/2 * 0.1573) = 5.72 x 10 -6 P(x | ψ b ) = 4 -1 * (5.72 x 10 -6 + 1.32 x 10 -6 + 1.58 x 10 -10 + 3.26 x 10 -6 ) = 2.57 x 10 -6
16
Method Overview
17
Foreground Probability: Foreground Modeling Foreground likelihood function
18
Foreground Modeling Foreground pixels: y[0]= [15, 15, 15, 1, 1]; Y[1]= [10,12,12.8,1,1]; Y[2]= [ 16,1,2,1,1]; Y[3]= [16, 10, 13, 1, 1]; Bandwidth matrix H: 16 0 0 0 0 0 16 0 0 0 0 0 16 0 0 0 0 0 31 0 0 0 0 0 21 Inverse of H: 0.063 0 0 0 0 0 0.063 0 0 0 0 0 0.063 0 0 0 0 0 0.032 0 0 0 0 0 0.048 |H| = 2666495 |H| -1/2 = 0.00061239 x = [16, 14, 15, 2, 1]; d = x – y[0] = [16-15,14-15,15-15,2-1,1-1] = [1,-1,0,1,0] d T H -1 d= 0.1573 Φ H = 0.00061239 * (2π) -5/2 * exp(-1/2 * 0.1573) = 6.188 x 10 -6 P(x | ψ f ) = (0.01 * (16*16*16*31*21) -1 ) + (1-0.01) * (4 -1 * 2.57 x 10 -6 ) Foreground Probability:
19
Method Overview
20
Likelihood Ratio Classifier
21
Spatial Context 4 Neighborhood Clique
22
Likelihood Function
23
Posterior
24
Log Posterior
26
Method Overview
27
Optimization: Graph Construction 2 1 5 4 3 S T τ 2 = 0.5 τ 1 = 0.2 -τ 3 = 0.07 -τ 4 = 0.01 -τ 5 = 0.1 1 1 1 1 1 1 11 1 1 1 τ 1 = 0.20 τ 2 = 0.50 τ 3 = -0.07 τ 4 = -0.01 τ 5 = -0.10 λ = 1 0.50 0.20-0.07-0.10 -0.01 Log Ratio Classifier for 4-Neighborhood Create a weighted graph G = {V, E}, where V = {v 1, v 2, v 2, v 4, v 5, s, t}, where s is the source and t is the sink. If τ i > 0, connect s (source) to v i with weigh τ i. Else, connect v i to t (sink) with weight -τ i. Next, add w(i, j) = λ if v i and v j are neighbors.
28
Why Minimum Cut? 2 1 5 4 3 S T τ 2 = 0.5 τ 1 = 0.2 -τ 3 = 0.07 -τ 4 = 0.01 -τ 5 = 0.1 1 1 1 1 1 1 11 1 1 1 τ 1 = 0.20 τ 2 = 0.50 τ 3 = -0.07 τ 4 = -0.01 τ 5 = -0.10 λ = 1 0.50 0.20-0.07-0.10 -0.01 Log Ratio Classifier for 4-Neighborhood The minimum cut corresponds to the max flow – and the weights of the max flow are equal to the parameters of the Log Posterior equation.
29
Method Overview
30
Model Update T = 23T = 24T = 25T = 26 T = 27T = 28 T = 29 T = next ρ b = 6 T = 26 T = 27T = 28 T = 29 T = next ρ f = 3
31
DEMO
32
All tests run on a 3.06 GHz Intel Pentium 4 with 1 GB RAM. Video sequences used a 240x360 resolution (0.08 megapixels). Bandwidth matrix H parameterized as a diagonal matrix with three equal variances for the range and two for the domain, with h r = 16 and h d = 25. Experimental Setup
33
Results
36
Mixture of Gaussians vs. Nonparametric
37
Object Level Detection Rates
38
Precision vs. Recall
39
Pixel-level detection recall and precision Using the Mixture of Gaussians
40
Uses a nonparametric kernel density estimator, which experimentally performs much better than a mixture of Gaussians estimator. Innovations include using the joint domain-range representation, which allows us to easily incorporate spatial distribution into the decision process. Also uses temporal persistence as a criterion for detection without feedback from higher level modules. All likelihoods calculated are used in a MAP-MRF framework to find an optimal global inference of the solution based on local information. Discussion
41
Weaknesses
42
Image stabilization – this algorithm only works for nominal camera motion Variant to frame rates, extremely fast moving and slow moving objects. Illumination Invariant. Future Work
43
Turlach, Berwin. “Bandwidth Selection in Kernel Density Estimation”. Dr. Gunturk's EE7750 Slides for Parameter Estimation http://vision.eecs.ucf.edu/projects/Detecting%20and%20Segmenting%20Hu mans%20in%20Crowded%20Scenes/detection_examples.jpg http://www.cs.ucf.edu/~sali/Projects/CoTrain/TitleImage.jpg http://www.philender.com/courses/multivariate/notes2/er9.gif http://math.bu.edu/people/sray/mat3.gif References
44
APPENDIX– Mixture of Gaussians
46
APPENDIX – Bandwidth Matrix Classes S positive scalar times the identity matrix D diagonal matrix with positive entries on the main diagonal F symmetric positive definite matrix
47
APPENDIX– Proper Kernel for KDE
48
Given: A doctor knows that meningitis causes stiff neck 50% of the time Prior probability of any patient having meningitis is 1/50,000 Prior probability of any patient having stiff neck is 1/20 If a patient has stiff neck, what’s the probability he/she has meningitis? APPENDIX : Example of Bayes Theorem
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.