 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.

Slides:



Advertisements
Similar presentations
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Advertisements

Supervised Learning Recap
Patch to the Future: Unsupervised Visual Prediction
Activity Recognition Aneeq Zia. Agenda What is activity recognition Typical methods used for action recognition “Evaluation of local spatio-temporal features.
Reducing Drift in Parametric Motion Tracking
(Includes references to Brian Clipp
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Oklahoma State University Generative Graphical Models for Maneuvering Object Tracking and Dynamics Analysis Xin Fan and Guoliang Fan Visual Computing and.
Hidden Markov Model based 2D Shape Classification Ninad Thakoor 1 and Jean Gao 2 1 Electrical Engineering, University of Texas at Arlington, TX-76013,
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Robust Video Stabilization Based on Particle Filter Tracking of Projected Camera Motion (IEEE 2009) Junlan Yang University of Illinois,Chicago.
Probabilistic video stabilization using Kalman filtering and mosaicking.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
MULTIPLE MOVING OBJECTS TRACKING FOR VIDEO SURVEILLANCE SYSTEMS.
Multi-camera Video Surveillance: Detection, Occlusion Handling, Tracking and Event Recognition Oytun Akman.
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
Isolated-Word Speech Recognition Using Hidden Markov Models
ECSE 6610 Pattern Recognition Professor Qiang Ji Spring, 2011.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Abstract Developing sign language applications for deaf people is extremely important, since it is difficult to communicate with people that are unfamiliar.
Learning and Recognizing Human Dynamics in Video Sequences Christoph Bregler Alvina Goh Reading group: 07/06/06.
A General Framework for Tracking Multiple People from a Moving Camera
earthobs.nr.no Temporal Analysis of Forest Cover Using a Hidden Markov Model Arnt-Børre Salberg and Øivind Due Trier Norwegian Computing Center.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Recognition, Analysis and Synthesis of Gesture Expressivity George Caridakis IVML-ICCS.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Using Inactivity to Detect Unusual behavior Presenter : Siang Wang Advisor : Dr. Yen - Ting Chen Date : Motion and video Computing, WMVC.
Maximum Entropy (ME) Maximum Entropy Markov Model (MEMM) Conditional Random Field (CRF)
Recognizing Action at a Distance Alexei A. Efros, Alexander C. Berg, Greg Mori, Jitendra Malik Computer Science Division, UC Berkeley Presented by Pundik.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
MURI Annual Review, Vanderbilt, Sep 8 th, 2009 Heterogeneous Sensor Webs for Automated Target Recognition and Tracking in Urban Terrain (W911NF )
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
A Dynamic Conditional Random Field Model for Object Segmentation in Image Sequences Duke University Machine Learning Group Presented by Qiuhua Liu March.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Hmm, HID HMMs Gerald Dalley MIT AI Lab Activity Perception Group Group Meeting 17 April 2003.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition Objectives: Reestimation Equations Continuous Distributions Gaussian Mixture Models EM Derivation of Reestimation Resources:
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Other Models for Time Series. The Hidden Markov Model (HMM)
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
Bayesian Decision Theory Case Studies CS479/679 Pattern Recognition Dr. George Bebis.
Particle Filtering for Geometric Active Contours
Dynamical Statistical Shape Priors for Level Set Based Tracking
Presented by: Yang Yu Spatiotemporal GMM for Background Subtraction with Superpixel Hierarchy Mingliang Chen, Xing Wei, Qingxiong.
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
LECTURE 15: REESTIMATION, EM AND MIXTURES
EM Algorithm and its Applications
Probabilistic Surrogate Models
Presentation transcript:

 Present by 陳群元

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 Tracking individuals in extremely crowded scenes is a challenging task,  we predict the local spatio-temporal motion patterns that describe the pedestrian movement at each space-time location in the video.  we robustly model the individual’s unique motion and appearance to discern them from surrounding pedestrians.

 Previous work track features and associate similar trajectories to detect individual moving entities within crowded scenes.  We encode many possible motions in the HMM, and derive a full distribution of the motion at each spatio-temporal location in the video.

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 An example : a 3-state Markov Chain λ o State 1 generates symbol A only, State 2 generates symbol B only, and State 3 generates symbol C only o Given a sequence of observed symbols O={CABBCABC}, the only one corresponding state sequence is {S 3 S 1 S 2 S 2 S 3 S 1 S 2 S 3 }, and the corresponding probability is P(O|λ)=P(q 0 =S 3 ) P(S 1 |S 3 )P(S 2 |S 1 )P(S 2 |S 2 )P(S 3 |S 2 )P(S 1 |S 3 )P(S 2 |S 1 )P(S 3 |S 2 ) =0.1  0.3  0.3  0.7  0.2  0.3  0.3  0.2= s2s2 s3s3 A B C s1s1

 An example : a 3-state discrete HMM λ o Given a sequence of observations O={ABC}, there are 27 possible corresponding state sequences, and therefore the corresponding probability is s2s2 s1s1 s3s3 {A:.3,B:.2,C:.5} {A:.7,B:.1,C:.2}{A:.3,B:.6,C:.1}

 f(Pos) = (f(Pos+1) -f(Pos) + f(Pos) -f(Pos-1))/2 = f(Pos+1)- f(Pos-1)/2;  For each pixel i in cuboid  I is intensity

 the local spatio-temporal motion pattern  represented by a 3D Gaussian of spatio-temporal gradients

 The hidden states of the HMM are represented by a set of motion patterns  The probability of an observed motion pattern given a hidden state s is

 Kullback–Leibler divergence is a non-symmetric measure of the difference between two probability distributions P and Q.

 After training a collection of HMMs on a video of typical crowd motion, we predict the motion pattern at each space- time location that contains the tracked subject.  where S is the set of hidden states, w(s) is defined by

Reference :A Tutorial On Hidden Markov Models andSelected Applications in Speech Recognition.

 a weighted sum of the 3D Gaussian distributions associated with the HMM’s hidden states

 The centroid we are interested in is a multivariate normal density that minimizes the total distortions. Formally, a centroid c is defined as, Reference: On Divergence Based Clustering of Normal Distributions and Its Application to HMM Adaptation

 where and are the mean and covariance of the hidden state s, respectively.

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 we use the gradient information to estimate the optical flow within each specific sub-volume and track the target in a Bayesian framework.  Bayesian tracking can be formulated as maximizing the posterior distribution of the state x t of the target at time t given available measurements z 1:t = {z i ; i = 1 : : : t} by  z t is the image at time t, p (x t |x t-1 ) is the transition distribution, and p (z t |x t ) is the likelihood.  state vector x t as the width, height, and 2D location of the target within the image.

 we focus on the target’s movement between frames and use a 2nd-degree autoregressive model for the transition distribution of the target’s width and height.  Ideally, the state transition distribution p (x t |x t-1 ) directly reflects the two-dimensional motion of the target between frames t -1 and t.  where is the 2D optical flow vector, and is the covariance matrix.

 Assuming the movement to be small, the image constraint at I(x,y,t) with Taylor series can be developed to get  H.O.T

 The predicted motion pattern is defined by a mean gradient vector and a covariance matrix  The motion information encoded in the spatio-temporal gradients can be expressed in the form of the structure tensor matrix  The optical flow can then be estimated from the structure tensor by solving  where w = [u; v; z] T is the 3D optical flow

 the 2D optical flow is v’ = [u/z; v/z]T

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 Typical models of the likelihood distribution p (z t |x t )  where is the variance, is a distance measure, and Z is a normalization term.  difference between a region R (defined by state x t ) of the observed image z t and the template.  We assume pedestrians exhibit consistency in their appearance and their motion, and model them in a joint likelihood by  where p A and p M are the appearance and motion likelihoods

 After tracking in frame t, we update each pixel i in the motion template by  where is the motion template at time t,  Is the region of spatio-temporal gradient defined by the tracking result (i.e., the expected value of the posterior)  is the learning rate.

 The error at pixel i and time t becomes  t i and r i are the normalized gradient vectors of the motion template and the tracking result at time t  To reduce the contributions of frequently changing pixels to the computation of the motion likelihood, we weigh each pixel in the likelihood’s distance measure.  where Z is a normalization term such that

 The distance measure of the motion likelihood distribution becomes

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 The training video for the concourse scene contains 300 frames (about 10 seconds of video),  the video for ticket gate scene contains 350 frames.  We set the cuboid size to 10*10*10 for both scenes.  The learning rate, appearance variance, and motion variance are 0.05.

 Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental results  conclusion

 In this paper, we derived a novel probabilistic method that exploits the inherent spatially and temporally varying structured pattern of a crowd’s motion to track individuals in extremely crowded scenes.

 The end  Thank you