Media IC & System Lab, GIEE, NTU Introduction to Augmented Reality

Slides:



Advertisements
Similar presentations
Distinctive Image Features from Scale-Invariant Keypoints
Advertisements

Automatic Color Gamut Calibration Cristobal Alvarez-Russell Michael Novitzky Phillip Marks.
Simultaneous surveillance camera calibration and foot-head homology estimation from human detection 1 Author : Micusic & Pajdla Presenter : Shiu, Jia-Hau.
QR Code Recognition Based On Image Processing
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
電腦視覺 Computer and Robot Vision I
BRISK (Presented by Josh Gleason)
Object Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition l Panoramas,
Computer vision: models, learning and inference
Mapping: Scaling Rotation Translation Warp
Chapter 6 Feature-based alignment Advanced Computer Vision.
Instructor: Mircea Nicolescu Lecture 13 CS 485 / 685 Computer Vision.
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
Image alignment Image from
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
A Study of Approaches for Object Recognition
Object Recognition with Invariant Features n Definition: Identify objects or scenes and determine their pose and model parameters n Applications l Industrial.
CS 223B Assignment 1 Help Session Dan Maynes-Aminzade.
Scale Invariant Feature Transform (SIFT)
Projected image of a cube. Classical Calibration.
MSU CSE 803 Fall 2008 Stockman1 CV: 3D sensing and calibration Coordinate system changes; perspective transformation; Stereo and structured light.
Camera Calibration CS485/685 Computer Vision Prof. Bebis.
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
Scale-Invariant Feature Transform (SIFT) Jinxiang Chai.
Automatic Camera Calibration
Computer vision: models, learning and inference
CSE 185 Introduction to Computer Vision
Shape Recognition and Pose Estimation for Mobile Augmented Reality Author : N. Hagbi, J. El-Sana, O. Bergig, and M. Billinghurst Date : Speaker.
1 Interest Operators Harris Corner Detector: the first and most basic interest operator Kadir Entropy Detector and its use in object recognition SIFT interest.
Object Tracking/Recognition using Invariant Local Features Applications l Mobile robots, driver assistance l Cell phone location or object recognition.
Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
3D SLAM for Omni-directional Camera
3D-2D registration Kazunori Umeda Chuo Univ., Japan CRV2010 Tutorial May 30, 2010.
Localization for Mobile Robot Using Monocular Vision Hyunsik Ahn Jan Tongmyong University.
CSCE 643 Computer Vision: Structure from Motion
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Distinctive Image Features from Scale-Invariant Keypoints David Lowe Presented by Tony X. Han March 11, 2008.
Reconnaissance d’objets et vision artificielle Jean Ponce Equipe-projet WILLOW ENS/INRIA/CNRS UMR 8548 Laboratoire.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
CSE 185 Introduction to Computer Vision Feature Matching.
A Tutorial on using SIFT Presented by Jimmy Huff (Slightly modified by Josiah Yoder for Winter )
3D Reconstruction Using Image Sequence
776 Computer Vision Jan-Michael Frahm Spring 2012.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Camera Calibration Course web page: vision.cis.udel.edu/cv March 24, 2003  Lecture 17.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Application of AR to Water System Isolation Task in NPP Md. Khalaquzzaman 7 th April 2008.
SIFT.
776 Computer Vision Jan-Michael Frahm Spring 2012.
Computer vision: models, learning and inference
SIFT Scale-Invariant Feature Transform David Lowe
Author : Sang Hwa Lee, Junyeong Choi, and Jong-Il Park
Presented by David Lee 3/20/2006
Distinctive Image Features from Scale-Invariant Keypoints
Paper – Stephen Se, David Lowe, Jim Little
Project 1: hybrid images
3D Vision Interest Points.
Fitting Curve Models to Edges
Structure from motion Input: Output: (Tomasi and Kanade)
From a presentation by Jimmy Huff Modified by Josiah Yoder
Uncalibrated Geometry & Stratification
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
SIFT.
Single-view geometry Odilon Redon, Cyclops, 1914.
CSE 185 Introduction to Computer Vision
The Pinhole Camera Model
Calibration and homographies
Structure from motion Input: Output: (Tomasi and Kanade)
Presentation transcript:

Media IC & System Lab, GIEE, NTU Introduction to Augmented Reality Po-Chen Wu Media IC & System Lab, GIEE, NTU Introduction to Augmented Reality Powerpoint Template: www.presentationmagazine.com Photo: www.circuitstoday.com, www.spinzomedia.com Texture: www.nipic.com

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu

Introduction R. T. Azuma[1] define an Augmented Reality (AR) system to have the following properties: Combines real and virtual Interactive in real time Registered in 3-D Video: https://www.youtube.com/watch?v=oH_LfXnklRw http://www.ad-dispatch.com/ [1] Azuma, Ronald T. "A survey of augmented reality." Presence 6.4 (1997): 355-385. Media IC & System Lab Po-Chen Wu

Marker-based AR Basic workflow of an AR application using fiducial marker tracking: Image: Wagner, Daniel, and Dieter Schmalstieg. Artoolkitplus for pose tracking on mobile devices. na, 2007. Media IC & System Lab Po-Chen Wu

Markerless AR Markerless augmented reality systems rely on natural features instead of fiducial marks. Image: http://www.arlab.com/ Media IC & System Lab Po-Chen Wu

Features Also known as interesting points, salient points or keypoints. Points that you can easily point out their correspondences in multiple images using only local information. Reference: Digital Visual Effects, Yung-Yu Chuang ? Media IC & System Lab Po-Chen Wu

Desired Properties for Features Distinctive: a single feature can be correctly matched with high probability. Invariant: invariant to scale, rotation, affine, illumination and noise for robust matching across a substantial range of affine distortion, viewpoint change and so on. That is, it is repeatable. Reference: Digital Visual Effects, Yung-Yu Chuang Media IC & System Lab Po-Chen Wu

Components Feature detection locates where they are. Feature description describes what they are. Feature matching decides whether two are the same one. Reference: Digital Visual Effects, Yung-Yu Chuang Media IC & System Lab Po-Chen Wu

Detectors & Descriptors Feature detector: Harris, Hessian, MSER, SIFT, SURF, FAST, etc. Feature descriptor: SIFT, SURF, DAISY, BRIEF, FREAK, etc. Reference: Digital Visual Effects, Yung-Yu Chuang Media IC & System Lab Po-Chen Wu

Another Markerless AR Complete registration with GPS, inertial sensors, and magnetic sensors. Image: www.theguardian.com Media IC & System Lab Po-Chen Wu

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu

System Flow Camera Calibration Marker Tracking Pose Estimation Rendering R,t Camera Coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) Marker Coordinates ( 𝑥 𝑐 , 𝑦 𝑐 ) Camera Screen Coordinates

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu

Camera Calibration (1/6) We use a simple cardboard frame with a ruled grid of lines for the camera calibration. Relationships between the camera screen coordinates and the camera coordinates can be known. Image: www.dreamincode.net Media IC & System Lab Po-Chen Wu

Camera Calibration (2/6) The relationships among the camera screen coordinates ( 𝑥 𝑐 , 𝑦 𝑐 ), the camera coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) and the marker coordinates ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) can be represented as below: ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 =𝐏 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 =𝐏∙ 𝐓 𝐜𝐦 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 =𝐂 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 = 𝐶 11 𝐶 21 𝐶 31 0 𝐶 12 𝐶 22 𝐶 32 0 𝐶 13 𝐶 23 𝐶 33 0 𝐶 14 𝐶 24 1 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 R,t Camera Coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) Marker Coordinates ( 𝑥 𝑐 , 𝑦 𝑐 ) Camera Screen Coordinates 𝐏= 𝑠 𝑥 𝑓 0 0 0 0 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 , 𝐓 𝐜𝐦 = 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 Intrinsic Matrix Extrinsic Matrix Media IC & System Lab Po-Chen Wu

Camera Calibration (3/6) The relationships between the camera screen coordinates ( 𝑥 𝑐 , 𝑦 𝑐 ) and the camera coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) can be represented as below: Intrinsic Matrix ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 =𝐏 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 = 𝑠 𝑥 𝑓 0 0 0 0 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 Camera Coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) Image: www.vaterrarc.com ( 𝑥 𝑐 , 𝑦 𝑐 ) Camera Screen Coordinates 𝑥 𝑐 = 𝑠 𝑥 𝑓 𝑋 𝑐 𝑍 𝑐 + 𝑥 0 𝑦 𝑐 = 𝑠 𝑦 𝑓 𝑌 𝑐 𝑍 𝑐 + 𝑦 0 Media IC & System Lab Po-Chen Wu

Camera Calibration (4/6) The relationships between the camera coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) and the marker coordinates ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) can be represented as below: 𝑋 𝑐 Marker Coordinates 𝑍 𝑚 Camera Coordinates Extrinsic Matrix 𝑋 𝑚 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 = 𝐓 𝐜𝐦 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 = 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 𝑌 𝑚 𝑍 𝑐 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 = 1 0 0 0 cos 𝜃 𝑥 − sin 𝜃 𝑥 0 sin 𝜃 𝑥 cos 𝜃 𝑥 cos 𝜃 𝑦 0 sin 𝜃 𝑦 0 1 0 −sin 𝜃 𝑦 0 cos 𝜃 𝑦 cos 𝜃 𝑧 − sin 𝜃 𝑧 0 sin 𝜃 𝑧 cos 𝜃 𝑧 0 0 0 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 + 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 𝑌 𝑐 Media IC & System Lab Po-Chen Wu

Camera Calibration (5/6) So the relationships between the camera screen coordinates ( 𝑥 𝑐 , 𝑦 𝑐 ) and the marker coordinates ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) can be represented as below: R,t Camera Coordinates ( 𝑋 𝑐 , 𝑌 𝑐 , 𝑍 𝑐 ) ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) Marker Coordinates ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 =𝐏 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 =𝐏∙ 𝐓 𝐜𝐦 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 =𝐂 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 = 𝐶 11 𝐶 21 𝐶 31 0 𝐶 12 𝐶 22 𝐶 32 0 𝐶 13 𝐶 23 𝐶 33 0 𝐶 14 𝐶 24 1 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 ( 𝑥 𝑐 , 𝑦 𝑐 ) Camera Screen Coordinates Image: www.andol.info 11 𝐏= 𝑠 𝑥 𝑓 0 0 0 0 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 , 𝐓 𝐜𝐦 = 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 4 6

Camera Calibration (6/6) A scalar variable 𝑘 is added into 𝐏 because matrix 𝐂 has 11 independent variables but matrices 𝐏 and 𝐓 𝐜𝐦 have 4 and 6 respectively. Since many pairs of ( 𝑥 𝑐 , 𝑦 𝑐 ) and ( 𝑋 𝑚 , 𝑌 𝑚 , 𝑍 𝑚 ) have been obtained by the procedure mentioned above, matrix 𝐂 can be estimated. ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 =𝐏 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 =𝐏∙ 𝐓 𝐜𝐦 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 =𝐂 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 = 𝐶 11 𝐶 21 𝐶 31 0 𝐶 12 𝐶 22 𝐶 32 0 𝐶 13 𝐶 23 𝐶 33 0 𝐶 14 𝐶 24 1 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 11 𝐏= 𝑠 𝑥 𝑓 0 0 0 𝑘 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 , 𝐓 𝐜𝐦 = 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 5 6

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Reference: Media IC & System Lab Po-Chen Wu

Marker Tracking The tracking pipeline consists of four basic steps[2]: Thresholding Fiducial Marker detection Rectangle fitting Pattern checking [2] Schmalstieg, Dieter, and Daniel Wagner. "Experiences with handheld augmented reality." ISMAR 2007. Media IC & System Lab Po-Chen Wu

Thresholding The first step in the image processing pipeline is to convert the input image into a binary image to reliably detect the black and white portions of the fiducial markers. A heuristic that has proven effective is to use the median of all marker pixels extracted in the last frame as a threshold for current image. Vignetting compensation incorporates the radial brightness into the per-pixel threshold value. Vignetting. Left: original camera image. Middle: constant thresholding. Right: thresholding with vignetting compensation.

Fiducial Marker Detection As a first step all scan-lines are searched left to right for edges. A sequence of white followed by black pixels is considered a candidate for a marker’s border. The software then follows this edge until either a loop back to the start pixel is closed or until the border of the image is reached. All pixels that have been visited are marked as processed in order to prevent following edges more than once. Left: Source image; Middle: Threshold image; Right: Three closed polygons as candidates for rectangle fitting.

Rectangle Fitting A first corner point c0 is selected as the contour point that lies at the maximum distance to an arbitrary starting point x of the contour. The center of the rectangle is estimated as the center of mass of all contour points. A line is formed from the first corner point and the center. Further corner points c1, c2 are found on each side of the line by searching for those points that have the largest distance to this line. These three corners c0, c1, c2 are used to construct more lines and recursively search for additional corners. An iterative process searches for corners until the whole contour has been searched or more than four corner points are detected. Example for fitting a rectangle to a polygon

Pattern Checking First a marker’s interior region is resampled into a normalized arrays of pixels. For perspectively correct unprojection, the homography matrix is computed from the markers’ corner points, which are assumed to form a rectangle. ARToolKit uses simple L2-norm template matching. There are several types of markers, and some are designed for efficiency. Marker types. Left: Template markers; Middle: BCHmarkers; Right: DataMatrix markers

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Reference: Kato, Hirokazu, and Mark Billinghurst. "Marker tracking and hmd calibration for a video-based augmented reality conferencing system," IWAR 1999. Media IC & System Lab Po-Chen Wu

Introduction to PnP Problem The aim of the Perspective-n-Point (PnP) problem is to determine the position and orientation of a camera given its: intrinsic parameters a set of n correspondences between 3D points and their 2D projections. (Xm, Ym, Zm) Marker Coordinates R,t Camera (Xc, Yc, Zc) (xc, yc) Camera Screen ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 = 𝑠 𝑥 𝑓 0 0 0 0 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 Pose

Papers for Solving PnP Problem "Fast and Globally Convergent Pose Estimation from Video Images," TPAMI 2000. "Complete Solution Classification for the Perspective-Three-Point Problem," TPAMI 2003. "Linear Pose Estimation from Points or Lines," TPAMI 2003. "Robust Pose Estimation from a Planar Target," TPAMI 2006. "Global Optimization through Searching Rotation Space and Optimal Estimation of the Essential Matrix," 2007 ICCV. "Globally Optimal O (n) Solution to the PnP Problem for General Camera Models," BMVC 2008. "EPnP An Accurate O(n) Solution to the PnP Problem," IJCV 2009. "Efficient lookup table based camera pose estimation for augmented reality, CAVW 2011. "A direct least-squares (DLS) method for PnP," ICCV 2011. "A Robust O(n) Solution to the Perspective-n-Point Problem," TPAMI 2012. "Revisiting the PnP Problem: A Fast, General and Optimal Solution," ICCV 2013. OI RPP EPnP RPnP OPnP Media IC & System Lab Po-Chen Wu

Position Estimation of Markers (1/5) All variables in the transformation matrix are determined by substituting screen coordinates and marker coordinates of detected marker's four vertices for ( 𝑥 𝑐 , 𝑦 𝑐 ) and ( 𝑋 𝑚 , 𝑌 𝑚 ) respectively[3]. When two parallel sides of a square marker are projected on the image, the equations of those line segments in the camera screen coordinates are the following ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ = 𝑁 11 𝑁 21 𝑁 31 𝑁 12 𝑁 22 𝑁 32 𝑁 13 𝑁 23 1 𝑋 𝑚 𝑌 𝑚 1 Perspective Transformation Matrix 𝑎 1 𝑥 𝑐 + 𝑏 1 𝑦 𝑐 + 𝑐 1 =0 𝑎 2 𝑥 𝑐 + 𝑏 2 𝑦 𝑐 + 𝑐 2 =0 [3] Kato, Hirokazu, and Mark Billinghurst. "Marker tracking and hmd calibration for a video-based augmented reality conferencing system," IWAR 1999.

Position Estimation of Markers (2/5) Given the intrinsic matrix 𝐏 that is obtained by the camera calibration, equations of the planes that include these two sides respectively can be represented in the camera coordinates. ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 =𝐏 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 = 𝑃 11 0 0 0 𝑃 12 𝑃 22 0 0 𝑃 13 𝑃 23 1 0 0 0 0 1 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 𝑎 1 𝑥 𝑐 + 𝑏 1 𝑦 𝑐 + 𝑐 1 =0 𝑎 2 𝑥 𝑐 + 𝑏 2 𝑦 𝑐 + 𝑐 2 =0 𝑎 1 𝑃 11 𝑋 𝑐 +( 𝑎 1 𝑃 12 + 𝑏 1 𝑃 22 ) 𝑌 𝑐 +( 𝑎 1 𝑃 13 + 𝑏 1 𝑃 23 + 𝑐 1 ) 𝑍 𝑐 =0 𝑎 2 𝑃 11 𝑋 𝑐 +( 𝑎 2 𝑃 12 + 𝑏 2 𝑃 22 ) 𝑌 𝑐 +( 𝑎 2 𝑃 13 + 𝑏 2 𝑃 23 + 𝑐 2 ) 𝑍 𝑐 =0

Position Estimation of Markers (3/5) Given that normal vectors of these planes are 𝐧 1 and 𝐧 2 respectively, the direction vector of parallel two sides of the square is given by the outer product 𝐮 1 = 𝐧 1 × 𝐧 2 . Given that two unit direction vectors that are obtained from two sets of two parallel sides of the square is 𝐮 1 and 𝐮 2 , these vectors should be perpendicular. 𝐮 1 𝐧 2 𝐮 2 𝐧 1 Media IC & System Lab Po-Chen Wu

Position Estimation of Markers (4/5) However, image processing errors mean that the vectors won't be exactly perpendicular. To compensate for this two perpendicular unit, direction vectors are defined by 𝐫 1 and 𝐫 2 in the plane that includes 𝐮 1 and 𝐮 2 . Given that the unit direction vector which is perpendicular to both 𝐫 1 and 𝐫 2 is 𝐫 3 , the rotation component 𝐑 3×3 in the transformation matrix 𝐓𝐜𝐦 from marker coordinates to camera coordinates is [ 𝐫 1 𝐫 2 𝐫 3 ]. 𝐫 1 𝐮 1 𝐫 1 𝐫 2 𝐫 3 𝑋 𝑐 𝑌 𝑐 𝑍 𝑐 1 = 𝐓 𝐜𝐦 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 = 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 𝐮 2 𝐫 2

Position Estimation of Markers (5/5) The four vertices coordinates of the marker in the marker coordinate frame and those coordinates in the camera screen coordinate frame, eight equations including translation component 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 are generated. The vertex coordinates of the markers in the marker coordinate frame can be transformed to coordinates in the camera screen coordinate frame by using the transformation matrix obtained. ℎ𝑥 𝑐 ℎ𝑦 𝑐 ℎ 1 = 𝑠 𝑥 𝑓 0 0 0 0 𝑠 𝑦 𝑓 0 0 𝑥 0 𝑦 0 1 0 0 0 0 1 𝑅 11 𝑅 21 𝑅 31 0 𝑅 12 𝑅 22 𝑅 32 0 𝑅 13 𝑅 23 𝑅 33 0 𝑇 𝑥 𝑇 𝑦 𝑇 𝑧 1 𝑋 𝑚 𝑌 𝑚 𝑍 𝑚 1 Media IC & System Lab Po-Chen Wu

Outline Introduction System Overview Camera Calibration Marker Tracking Pose Estimation of Markers Conclusion Media IC & System Lab Po-Chen Wu

Conclusion We propose a method for tracking fiducial markers and a calibration method for camera based on the marker tracking. The most important thing is to get the intrinsic and extrinsic matrix. The rendering part can be implemented with the transformation matrix obtained. Topics of GPU group. Media IC & System Lab Po-Chen Wu