Probabilistic video stabilization using Kalman filtering and mosaicking.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Visual Servo Control Tutorial Part 1: Basic Approaches Chayatat Ratanasawanya December 2, 2009 Ref: Article by Francois Chaumette & Seth Hutchinson.
Reducing Drift in Parametric Motion Tracking
Mapping: Scaling Rotation Translation Warp
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Yen-Lin Lee and Truong Nguyen ECE Dept., UCSD, La Jolla, CA Method and Architecture Design for Motion Compensated Frame Interpolation in High-Definition.
Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes (CVPR’10) Shengcai Liao, Guoying Zhao, Vili Kellokumpu,
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
1 Robust Video Stabilization Based on Particle Filter Tracking of Projected Camera Motion (IEEE 2009) Junlan Yang University of Illinois,Chicago.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
Adam Rachmielowski 615 Project: Real-time monocular vision-based SLAM.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Optical Flow Methods 2007/8/9.
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
Detecting and Tracking Moving Objects for Video Surveillance Isaac Cohen and Gerard Medioni University of Southern California.
Static Image Mosaicing
Object Detection and Tracking Mike Knowles 11 th January 2005
Tracking using the Kalman Filter. Point Tracking Estimate the location of a given point along a sequence of images. (x 0,y 0 ) (x n,y n )
Discriminative Training of Kalman Filters P. Abbeel, A. Coates, M
Effective Gaussian mixture learning for video background subtraction Dar-Shyang Lee, Member, IEEE.
Augmented Reality: Object Tracking and Active Appearance Model
© 2003 by Davi GeigerComputer Vision November 2003 L1.1 Tracking We are given a contour   with coordinates   ={x 1, x 2, …, x N } at the initial frame.
Tracking with Linear Dynamic Models. Introduction Tracking is the problem of generating an inference about the motion of an object given a sequence of.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Hand Signals Recognition from Video Using 3D Motion Capture Archive Tai-Peng Tian Stan Sclaroff Computer Science Department B OSTON U NIVERSITY I. Introduction.
Digital Image Stabilization 老師 : 楊士萱 學生 : 鄭馥銘. Outline Introduction Basic architecture of DIS MVI method for DIS Future work.
Overview and Mathematics Bjoern Griesbach
Optical flow (motion vector) computation Course: Computer Graphics and Image Processing Semester:Fall 2002 Presenter:Nilesh Ghubade
Adaptive Signal Processing Class Project Adaptive Interacting Multiple Model Technique for Tracking Maneuvering Targets Viji Paul, Sahay Shishir Brijendra,
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
A plane-plus-parallax algorithm Basic Model: When FOV is not very large and the camera motion has a small rotation, the 2D displacement (u,v) of an image.
1 Formation et Analyse d’Images Session 7 Daniela Hall 7 November 2005.
Kalman filter and SLAM problem
Muhammad Moeen YaqoobPage 1 Moment-Matching Trackers for Difficult Targets Muhammad Moeen Yaqoob Supervisor: Professor Richard Vinter.
HMM-BASED PSEUDO-CLEAN SPEECH SYNTHESIS FOR SPLICE ALGORITHM Jun Du, Yu Hu, Li-Rong Dai, Ren-Hua Wang Wen-Yi Chu Department of Computer Science & Information.
Tracking Pedestrians Using Local Spatio- Temporal Motion Patterns in Extremely Crowded Scenes Louis Kratz and Ko Nishino IEEE TRANSACTIONS ON PATTERN ANALYSIS.
A Shaft Sensorless Control for PMSM Using Direct Neural Network Adaptive Observer Authors: Guo Qingding Luo Ruifu Wang Limei IEEE IECON 22 nd International.
Multimodal Interaction Dr. Mike Spann
Complete Pose Determination for Low Altitude Unmanned Aerial Vehicle Using Stereo Vision Luke K. Wang, Shan-Chih Hsieh, Eden C.-W. Hsueh 1 Fei-Bin Hsaio.
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
1 University of Texas at Austin Machine Learning Group 图像与视频处理 计算机学院 Motion Detection and Estimation.
Advances in digital image compression techniques Guojun Lu, Computer Communications, Vol. 16, No. 4, Apr, 1993, pp
Real-Time Simultaneous Localization and Mapping with a Single Camera (Mono SLAM) Young Ki Baik Computer Vision Lab. Seoul National University.
Optical Flow. Distribution of apparent velocities of movement of brightness pattern in an image.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Range-Only SLAM for Robots Operating Cooperatively with Sensor Networks Authors: Joseph Djugash, Sanjiv Singh, George Kantor and Wei Zhang Reading Assignment.
Multimodal Interaction Dr. Mike Spann
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Tracking with dynamics
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Representing Moving Images with Layers J. Y. Wang and E. H. Adelson MIT Media Lab.
MOTION Model. Road Map Motion Model Non Parametric Motion Field : Algorithms 1.Optical flow field estimation. 2.Block based motion estimation. 3.Pel –recursive.
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
Tracking We are given a contour G1 with coordinates G1={x1 , x2 , … , xN} at the initial frame t=1, were the image is It=1 . We are interested in tracking.
Motion Detection And Analysis
The Brightness Constraint
Dynamical Statistical Shape Priors for Level Set Based Tracking
Range Imaging Through Triangulation
The Brightness Constraint
Exposing Digital Forgeries by Detecting Traces of Resampling Alin C
The Brightness Constraint
Image and Video Processing
Image and Video Processing
NONLINEAR AND ADAPTIVE SIGNAL ESTIMATION
NONLINEAR AND ADAPTIVE SIGNAL ESTIMATION
Presentation transcript:

Probabilistic video stabilization using Kalman filtering and mosaicking

ABSTRACT  The removal of unwanted, parasitic vibrations in a video sequence induced by camera motion is an essential part of video acquisition.  We present a new image processing method to remove such vibrations and reconstruct a video sequence void of sudden camera movement.

INTRODUCTION-1  An approach (optical stabilization) consists of implementing an optical system that compensates for unwanted camera motion using motion sensor and active optical system.  The most powerful, but makes video cameras significantly expensive.

INTRODUCTION-2  This paper is focus on another approach consists in performing post-processing of the video sequence to eliminate unwanted motion in the video (swings and twists) caused by a person holding the camera or mechanical vibration.

VIDEO STABILIZATION AND RECONSTRUCT FRAMEWORK-1  The overall algorithm consists of the following steps: 1.Video sequence stabilization 1)Estimation of the pair-wise transformations between adjacent frames. 2)Estimation of the intentional motion parameters (Kalman filtering in time). 3)Compensation of each frame for unwanted motion (frame warping).

VIDEO STABILIZATION AND RECONSTRUCT FRAMEWORK-2 2.Reconstruction of undefined regions using mosaicking: 1)Estimation of the transformation between distant frame. 2)Warping distant frames and constructing mosaic for undefined regions in each frame.

The block diagram of overall algorithm

Estimation of the pair-wise transformations between adjacent frames-1  Under an affine transformation, pixel locations in frames and are related by a transformation given by where and are pixel coordinates before and after transformation respectively. elements of matrix A describe zoom, rotation and dolly motion of the camera, and vector b describes panning and tracking motion.

Estimation of the pair-wise transformations between adjacent frames-2  Transform of,aligning frames and,is estimated by minimizing the following cost function with respect to where m=1 and is the set of all locations in the image plane for which transformed coordinates lie in the limits of the valid image coordinate.  The choice of function φ(x) is crucial for robustness of the transformation.

Estimation of the pair-wise transformations between adjacent frames-3  Here we use an approximation to the -norm given by with β=0.01,insures differentiability of the cost function near zero; p=1 chosen empirically from several test sequences…  In order to avoid local minima of the cost function (2) and accelerate the convergence we use a multi-scale implement.

Estimating intentional motion parameters-1  The cumulative transform for frame n, denoted by,can be obtained as follows: Elements of matrix A describe zoom, rotation and dolly motion of the camera, and vector b describes panning and tracking motion.  Similarly, we describe the image transform parameters representing intentional motion in terms of intentional cumulative transform

Estimating intentional motion parameters-2  Optimal estimation of is carry out using a recursive Kalman filtering algorithm.  We treat as noisy observations of intentional cumulative transform parameters obeying physics-based dynamic model.  The state model for each of the parameters depends on real-life expected behavior of these parameters.  Two distinct behavior patterns can be identified leading to different dynamic models for different parameters.

Two distinct behavior patterns leading to different dynamic models for different parameters  We introduce velocity variables for each,respectively.  It is reasonable to assume the independence of dynamic models for each of the 4 parameters. For example, and follow the dynamic model given by where is white Gaussian noise with variance.

Two distinct behavior patterns leading to different dynamic models for different parameters  The remaining parameters and are assumed to be constant in the absence of noise.  The simple dynamic model for and we have (for )

 The overall state-space model for the intentional cumulative transform parameters is given by variance of the noise term is different for each kind of variable.

 The observed cumulative transform parameters are treated as noisy observations of the intentional cumulative transform parameters.  The observation model for each parameters is independent, leading to observation model  Observation noise variances describe the variability of unwanted transformation between frames.

Compensation of each frame for unwanted transformation (frame warping)  Resulting transform is given by where and are initial and transformed coordinates in frame n.  Using (9), a warped frame is computed as follows p.s. computing image values at non- integer locations in (10) is carried out by cubic interpolation.

Reconstruction of undefined regions using mosaicking  After compensating transformation is applied to each frame, undefined regions appear near the edge of each frame.  The extent of these regions varies from frame to frame and presents unacceptable visual artifacts.  Use frame trimming and magnification or filling by a constant value, lead to severe quality degradation of the resulting video and limit the range of possible correcting transformations.  Here we propose to use mosaicking for each frame in order to exploit temporal correlations between frames.

Mosaicking illustration

Estimation of the transformation between distant frames-1  In order to properly align up to M future and past frames with respect to the current warped frame n, we need to find registration parameters of these frames with respect to the current frame.  For a given frame n, as initial conditions, we sequentially estimate the global transform parameters between frame n and n ± m, where 2≤ m ≤M.  For each m, cascaded transforms and are used to initialize the solution for.

Estimation of the transformation between distant frames-2  The coordinate transformation obtained using cascaded transforms is given by  For instance, transformation for past frame (-m) with respect to frame n can be found by inverting the registration of frame n as a future frame (n - m)

Warping distant frames and composing mosaic for regions in each frame-1  Each frame out of 2M neighboring frames is aligned with respect to the warped current frame given by (10).  Aligning transform for frame is formed by cascading inverted registration transform with the correcting transform defined in (9).  The resulting warping transform is given by

Warping distant frames and composing mosaic for regions in each frame-2  And the warped frame is computed as follows  For each undefined pixel x in the target frame,the reconstructed image value is found as follows where the weights are set to the inverse of the errors of registrationobtained by minimizing (2).

RESULTS  We test our technique on 3 real-life video sequences (which we call A, B and C). ABCABC  To simplify the task we modify the motion model, first we assume only translational motion between described by vector b.  Using this model, the cumulative transform parameters are given by the components of for sequence A are shown in Figure3.

Cumulative motion parameters for sequence A

 Assume static camera (performing “ total motion compensation ” ), the correcting transform becomes the result of applying such compensating transform is illustrated in Figure4.

 In the figure, it can be seen that landmark objects in the corrected sequence do not move with respect to the frame coordinates, while rotational vibrations remain uncorrected. Shift A A

Full 6-parameter inter-frame affine motion model

Full result of A, B, and C. ABCABC

CONCLUSIONS  Using our technique we obtained promising preliminary results on random test sequences with complex motion and severe vibrations.  We compared our results with one of commercial products and showed a significant improvement of performance for our technique. Compared A and B. ABAB  Our method of stabilization can be easily adapted to perform additional processing, such as sampling rate conversion, static mosaic construction, ego-motion estimation.