Augmented Reality: Object Tracking and Active Appearance Model

Slides:



Advertisements
Similar presentations
Active Appearance Models
Advertisements

Face Alignment by Explicit Shape Regression
Motion.
DDDAS: Stochastic Multicue Tracking of Objects with Many Degrees of Freedom PIs: D. Metaxas, A. Elgammal and V. Pavlovic Dept of CS, Rutgers University.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
Face Alignment with Part-Based Modeling
AAM based Face Tracking with Temporal Matching and Face Segmentation Dalong Du.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Computer Vision Optical Flow
Virtual Dart: An Augmented Reality Game on Mobile Device Supervisor: Professor Michael R. Lyu Prepared by: Lai Chung Sum Siu Ho Tung.
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
A New Block Based Motion Estimation with True Region Motion Field Jozef Huska & Peter Kulla EUROCON 2007 The International Conference on “Computer as a.
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Motion Detection And Analysis Michael Knowles Tuesday 13 th January 2004.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
A Study of Approaches for Object Recognition
Optical Flow Methods 2007/8/9.
1 Static Sprite Generation Prof ︰ David, Lin Student ︰ Jang-Ta, Jiang
Probabilistic video stabilization using Kalman filtering and mosaicking.
Rodent Behavior Analysis Tom Henderson Vision Based Behavior Analysis Universitaet Karlsruhe (TH) 12 November /9.
Real-time Combined 2D+3D Active Appearance Models Jing Xiao, Simon Baker,Iain Matthew, and Takeo Kanade CVPR 2004 Presented by Pat Chan 23/11/2004.
Feature matching and tracking Class 5 Read Section 4.1 of course notes Read Shi and Tomasi’s paper on.
Computing motion between images
ART: Augmented Reality Table for Interactive Trading Card Game Albert H.T. Lam, Kevin C. H. Chow, Edward H. H. Yau and Michael R. Lyu Department of Computer.
Motion Computing in Image Analysis
Tracking Video Objects in Cluttered Background
Multiple View Geometry Marc Pollefeys University of North Carolina at Chapel Hill Modified by Philippos Mordohai.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Numerical Recipes (Newton-Raphson), 9.4 (first.
Presented by Pat Chan Pik Wah 28/04/2005 Qualifying Examination
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Jacinto C. Nascimento, Member, IEEE, and Jorge S. Marques
Overview and Mathematics Bjoern Griesbach
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
1 Activity and Motion Detection in Videos Longin Jan Latecki and Roland Miezianko, Temple University Dragoljub Pokrajac, Delaware State University Dover,
Computer vision: models, learning and inference
Shape Recognition and Pose Estimation for Mobile Augmented Reality Author : N. Hagbi, J. El-Sana, O. Bergig, and M. Billinghurst Date : Speaker.
TP15 - Tracking Computer Vision, FCUP, 2013 Miguel Coimbra Slides by Prof. Kristen Grauman.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Detecting Pedestrians Using Patterns of Motion and Appearance Paul Viola Microsoft Research Irfan Ullah Dept. of Info. and Comm. Engr. Myongji University.
High-Resolution Interactive Panoramas with MPEG-4 발표자 : 김영백 임베디드시스템연구실.
The Measurement of Visual Motion P. Anandan Microsoft Research.
CSCE 643 Computer Vision: Structure from Motion
Digital Image Processing Lecture 6: Image Geometry
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
Tracking People by Learning Their Appearance Deva Ramanan David A. Forsuth Andrew Zisserman.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
3D Reconstruction Using Image Sequence
Representing Moving Images with Layers J. Y. Wang and E. H. Adelson MIT Media Lab.
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Motion and Optical Flow
Image Primitives and Correspondence
Range Imaging Through Triangulation
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
CSSE463: Image Recognition Day 30
Announcements Questions on the project? New turn-in info online
Coupled Horn-Schunck and Lukas-Kanade for image processing
Detection of salient points
CSSE463: Image Recognition Day 30
Optical flow and keypoint tracking
Presentation transcript:

Augmented Reality: Object Tracking and Active Appearance Model Presented by Pat Chan 01/03/2005 Group Meeting

Outline Introduction to Augmented Reality Object Tracking Active Appearance Model (AAM) Object Tracking with AAM Future Direction Conclusion

Introduction An Augmented Reality system supplements the real world with virtual objects that appear to coexist in the same space as the real world Properties : Combine real and virtual objects in a real environment Runs interactively, and in real time Registers(aligns) real and virtual objects with each other

Introduction Display Tracking 3D Modeling Registration Presenting virtual objects on real environment Tracking Following user’s and virtual object’s movements by means of a special device or techniques 3D Modeling Forming virtual object Registration Blending real and virtual objects

Object Tracking Visual content can be modeled as a hierarchy of abstractions. At the first level are the raw pixels with color or brightness information. Further processing yields features such as edges, corners, lines, curves, and color regions. A higher abstraction layer may combine and interpret these features as objects and their attributes.

Object Tracking Accurately tracking the user’s position is crucial for AR registration The objective is to obtain an accurate estimate of the position (x,y) of the object tracked Tracking = correspondence + constraints + estimation Tracking objects is a sequence of video frames is composed of two main stages: Isolation of objects from background in each frames Association of objects in successive frames in order to trace them For prepared indoor environments, systems employ hybrid-tracking technique such as magnetic and video sensors to exploit strengths and compensate weaknesses of individual tracking technologies. In outdoor and mobile AR application, it generally isn’t practical to cover the environment with markers. Network-based tracking method for Indoor/outdoor

Object Tracking Object Tracking in image processing is usually based on reference image of the object, or properties of the objects. Tracking techniques: Kalman filtering Correlation-based tracking, Change-based tracking 2D layer tracking tracking of articulated objects

Object Tracking Object Tracking can be briefly divides into following stages: Input (object and camera) Finding correspondence Motion Estimation Corrective Feedback Occlusion Detection

Input Tracking algorithms can be classified into Single object & Single Camera Single object & Multiple Cameras Multiple object & Single Camera Multiple objects & Multiple Cameras

Single Object & Single Camera Accurate camera calibration and scene model Suffers from Occlusions Not robust and object dependant

Single Object & Multiple Camera Accurate point correspondence between scenes Occlusions can be minimized or even avoided Redundant information for better estimation Multiple camera Communication problem

Possible Solution

Static Point Correspondence The output of the tracking stage is A simple scene model is used to get real estimation of coordinates Both Affine and Perspective models were used for the scene modeling Static corresponding points were used for parameter estimation Least mean squares was used to improve parameter estimation

Dynamic Point Correspondence

Block-Based Motion Estimation Typically, in object tracking precise sub-pixel optical flow estimation is not needed. Motion can be in the order of several pixels, thereby precluding use of gradient methods. A simple sum of squared differences error criterion coupled with full search in a limited region around the tracking window can be applied.

Adaptive Window Sizing Although simple block-based motion estimation may work reasonably well when motion is purely translational It can lose the object if its relative size changes. If the object’s camera field of view shrinks, the SSD error is strongly influenced by the background. If the object’s camera field of view grows, the window fails to make use of entire object information and can slip away.

Four Corner Method This technique divides the rectangular object window into 4 basic regions - each one quadrant. Motion vectors are calculated for each subregion and each controls one of four corners. Translational motion is captured by all four moving equally, while window size is modulated when motion is differential. Resultant tracking window can be non-rectangular, i.e., any quadrilateral approximated by four rectangles with a shared center corner.

Example: Four Corner Method Synthetically generated test sequences:

Correlative Method Four corner method is strongly subject to error accumulation which can result in drift of one or more of the tracking window quadrants. Once drift occurs, sizing of window is highly inaccurate. Need a method that has some corrective feedback so window can converge to correct size even after some errors. Correlation of current object features to some template view is one solution.

Correlative Method (con’t) Basic form of technique involves storing initial view of object as a reference image. Block matching is performed through a combined interframe and correlative MSE: where sc’(x0,y0,0) is the resized stored template image. Furthermore, minimum correlative MSE is used to direct resizing of current window.

Example: Correlative Method

Occlusion Detection Each camera must possess an ability to assess the validity of its tracking (e.g. to detect occlusion). Comparing the minimum error at each point to some absolute threshold is problematic since error can grow even when tracking is still valid. Threshold must be adaptive to current conditions. One solution is to use a threshold of k (constant > 1) times the moving average of the MSE. Thus, only steep changes in error trigger indication of possibly wrong tracking.

Improvements Things can be improved Good filtering algorithms Adequate dynamical models Shape/appearance models need work

Active Appearance Models (AAMs) Active Appearance Models are generative models commonly used to model faces Can also be useful for other phenomena Matching object classes Deformable appearance models Another closely related type of face models are 3D Morphable Models In this paper, it tries to model 3D phenomena by using the 2D AAM Constrain the AAM with the 3D models to achieve a real-time algorithm for fitting the AAM

Active Appearance Models (AAMs) 2D linear shape is defined by 2D triangulated mesh and in particular the vertex locations of the mesh. Shape s can be expressed as a base shape s0. pi are the shape parameter. s0 is the mean shape and the matrices si are the eigenvectors corresponding to the m largest eigenvalues 68 vertices

Active Appearance Models (AAMs) The appearance of an independent AAM is defined within the base mesh s0. A(u) defined over the pixels u ∈ s0 A(u) can be expressed as a base appearance A0(u) plus a linear combination of l appearance Coefficients λi are the appearance parameters. A0(u) A1(u) A2(u) A3(u)

Active Appearance Models (AAMs) The AAM model instance with shape parameters p and appearance parameters λ is then created by warping the appearance A from the base mesh s0 to the model shape s. Piecewise affine warp W(u; p): (1) for any pixel u in s0 find out which triangle it lies in, (2) warp u with the affine warp for that triangle. M(W(u;p))

Fitting AAMs Minimize the error between I (u) and M(W(u; p)) = A(u). If u is a pixel in s0, then the corresponding pixel in the input image I is W(u; p). At pixel u the AAM has the appearance At pixel W(u; p), the input image has the intensity I (W(u; p)). Minimize the sum of squares of the difference between these two quantities: 1. For each pixel x in the base mesh s0, we compute the corresponding pixel W(x; p) in the input image by warping x with the piecewise affine warp W. 2. The input image I is then sampled at the pixel W(x; p); typically it is bi-linearly interpolated at this pixel. 3. The resulting value is then subtracted from the appearance at that pixel and the result stored in E u

Object Tracking with AAM Objects can be tracked with the trained AAM 3-D face tracking with AAM search Pose estimation with AAM

Example aam_tracking_mpeg4.avi The training set consisted of five images of a DAT tape cassette DAT cassette was annotated using 12 landmarks Upon the five training images, a two-level multi-scale AAM was built. aam_tracking_mpeg4.avi

Future Direction Propose a general object tracking algorithm with the help of AAM Improve the accuracy of the object tracking algorithm Improve the fitting speed of the AAM

Conclusion Introduction on Augmented Reality Survey on Object Tracking Introduction Active Appearance Model Improve the accuracy of object tracking by AAM Proposed our future research direction