CMSC5711 Image processing and computer vision

Slides:



Advertisements
Similar presentations
Geometry of Aerial Photographs
Advertisements

Single-view geometry Odilon Redon, Cyclops, 1914.
Image Processing and Computer Vision Chapter 10: Pose estimation by the iterative method (restart at week 10) Pose estimation V4h31.
Computer Vision, Robert Pless
QR Code Recognition Based On Image Processing
Last 4 lectures Camera Structure HDR Image Filtering Image Transform.
Dynamic Occlusion Analysis in Optical Flow Fields
MASKS © 2004 Invitation to 3D vision Lecture 7 Step-by-Step Model Buidling.
Camera Models A camera is a mapping between the 3D world and a 2D image The principal camera of interest is central projection.
Motion Tracking. Image Processing and Computer Vision: 82 Introduction Finding how objects have moved in an image sequence Movement in space Movement.
Lecture 5: Projection CS6670: Computer Vision Noah Snavely.
CS485/685 Computer Vision Prof. George Bebis
Vision, Video and Virtual Reality 3D Vision Lecture 12 Camera Models CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
3D Computer Vision and Video Computing 3D Vision Lecture 14 Stereo Vision (I) CSC 59866CD Fall 2004 Zhigang Zhu, NAC 8/203A
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Camera parameters Extrinisic parameters define location and orientation of camera reference frame with respect to world frame Intrinsic parameters define.
CSE 6367 Computer Vision Stereo Reconstruction Camera Coordinate Transformations “Everything should be made as simple as possible, but not simpler.” Albert.
Automatic Camera Calibration
Lecture 14: Projection CS4670 / 5670: Computer Vision Noah Snavely “The School of Athens,” Raphael.
WP3 - 3D reprojection Goal: reproject 2D ball positions from both cameras into 3D space Inputs: – 2D ball positions estimated by WP2 – 2D table positions.
Image Formation. Input - Digital Images Intensity Images – encoding of light intensity Range Images – encoding of shape and distance They are both a 2-D.
Course 12 Calibration. 1.Introduction In theoretic discussions, we have assumed: Camera is located at the origin of coordinate system of scene.
Geometric Models & Camera Calibration
3D-2D registration Kazunori Umeda Chuo Univ., Japan CRV2010 Tutorial May 30, 2010.
Localization for Mobile Robot Using Monocular Vision Hyunsik Ahn Jan Tongmyong University.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
CSCE 643 Computer Vision: Structure from Motion
Correspondence-Free Determination of the Affine Fundamental Matrix (Tue) Young Ki Baik, Computer Vision Lab.
Geometric Camera Models
Binocular Stereo #1. Topics 1. Principle 2. binocular stereo basic equation 3. epipolar line 4. features and strategies for matching.
Computer Vision : CISC 4/689 Going Back a little Cameras.ppt.
Copyright Howie Choset, Renata Melamud, Al Costa, Vincent Lee-Shue, Sean Piper, Ryan de Jonckheere. All Rights Reserved Computer Vision.
CSC508 Convolution Operators. CSC508 Convolution Arguably the most fundamental operation of computer vision It’s a neighborhood operator –Similar to the.
The geometry of the system consisting of the hyperbolic mirror and the CCD camera is shown to the right. The points on the mirror surface can be expressed.
1 Chapter 2: Geometric Camera Models Objective: Formulate the geometrical relationships between image and scene measurements Scene: a 3-D function, g(x,y,z)
October 16, 2014Computer Vision Lecture 12: Image Segmentation II 1 Hough Transform The Hough transform is a very general technique for feature detection.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Robotics Chapter 6 – Machine Vision Dr. Amit Goradia.
Instructor: Mircea Nicolescu Lecture 7
Problem Set 2 Reconstructing a Simpler World COS429 Computer Vision Due October (one week from today)13 th.
Lecture 14: Projection CS4670 / 5670: Computer Vision Noah Snavely “The School of Athens,” Raphael.
3D Computer Vision 3D Vision Lecture 3 - Part 1 Camera Models Spring 2006 Zhang Aiwu.
Recognizing specific objects Matching with SIFT Original suggestion Lowe, 1999,2004.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae & david streader, VUW Images and 2D Graphics COMP.
Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.
CMSC5711 Image processing and computer vision
Images and 2D Graphics COMP
CS4670 / 5670: Computer Vision Kavita Bala Lecture 20: Panoramas.
Paper – Stephen Se, David Lowe, Jim Little
Digital Visual Effects, Spring 2007 Yung-Yu Chuang 2007/4/17
CSCE 441 Computer Graphics 3-D Viewing
Mean Shift Segmentation
CMSC5711 Revision (1) (v.7.b) revised
CMSC5711 Revision 3 CMSC5711 revision 3 ver.x67.8c.
WINDOWING AND CLIPPING
Lecture 3: Camera Rotations and Homographies
Chapter 1: Image processing and computer vision Introduction
Image processing and computer vision
Image Processing, Lecture #8
The SIFT (Scale Invariant Feature Transform) Detector and Descriptor
WINDOWING AND CLIPPING
Image Processing, Lecture #8
Image processing and computer vision
Computer and Robot Vision I
CS 565 Computer Vision Nazar Khan Lecture 9.
Computer and Robot Vision I
Computer and Robot Vision I
Revision 4 CSMSC5711 Revision 4: CSMC5711 v.9a.
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

CMSC5711 Image processing and computer vision Revision 2 CMSC5711 Revision 1 V.x56.7b

Q2.1 A 3D point M is at [X,Y,Z]T=[0.2, 0.1, 2.0] T (in meters) in the world coordinate system. The parameters of a camera are shown below: The focal length F=3mm. Horizontal pixel width sx= Vertical pixel width sy=5m. The CCD sensor size = 10 mm x 10 mm. The image centre is at the centre of the image plane. The origin (1,1) of the image is at the right-bottom corner. The x-coordinate is increasing from right to left, and y-coordinate is increasing from bottom to top. You may assume the camera coordinates are the same as the world coordinates. Estimate the size (Xmax, Ymax) of the image captured by the camera in pixels. Estimate the image centre (Ox, Oy) in pixels measuring from the right-bottom corner of the CCD sensor. Find the focal length in pixels. Find the 3-D position of the point M in pixels. M is being translated by T= [1,2,3]T meters first and then rotated around a point P=[1.5, 2.0, 5.5] T, the rotation angles are [ 1 , 2 , 3 ]=[0.1, 0.2, 0.3] in degrees. Find the new 3-D position (M’) of M in pixels. Find the 2-D image position (in pixels) of M’. CMSC5711 Revision 1 V.x56.7b

Q2.2 Convolution , edge mask and edge detection An image S and a mask m are shown below   Find the convolution result of S and m, the result should include all partially overlapping cases. Show the matrix of a 3 x 3 second order edge mask that you know. How do you find an edge image using a second order edge mask? CMSC5711 Revision 1 V.x56.7b

An original gray level image has resolution M=50 rows and N=50 columns An original gray level image has resolution M=50 rows and N=50 columns. The gray level resolution (L) of each pixel is 8 (gray level from 0 to 7). R(k) is the gray level of index k, N(k) is the number of pixels that have gray level R(k). Pr(R(k)) is the probability of the pixels in the image having gray level R(k). After histogram normalization, S(k) is the normalized gray level of index k. A table to help you to perform histogram equalization is shown below. Find the value of x in the table. Discuss the relation between pixel resolution (bits per pixel) and the result of histogram normalization in image processing. For the following table (you may copy it to your answer book first), fill in the blanks. Discuss how to use histogram equalization to make a colour picture looks better. Q2.3 r(k) N(k) Pr(r(k)) S(k) Round off (S(k)) r(0) 13   r(1) 46 r(2) 285 r(3) 645 r(4) 777 r(5) 490 r(6) 91 r(7) x CMSC5711 Revision 1 V.x56.7b

Q2.4 Structure from motion Initially at time = 0, the centre of all (total four, index j=1,2,3,4) image feature points of an object is at [0,0]T. At time =t, the object is moved to a new position and the projected image points x(j,t) (in pixels) are at x(j=1,t)=(a,b) x(j=2,t)=(c,d) x(j=3,t)=(e,f) x(j=4,t)=(g,h) (a) If an orthographic camera model is used, calculate the image translation of the object from time =0 to time =t in terms of a,b,c,d,e,f,g and h.   (b) The projection matrix of a camera is of size 3 by 4. Discuss the constraints for the values in the projection matrix for each of the following camera models : A perspective camera. An affine camera. An orthographic camera. (c) Discuss the relative advantage and disadvantages of using (i) factorization and (ii) bundle adjustment for finding the structure and pose of an object from an image sequence. You may comment on their speeds, accuracies and the types of approaches (linear or iterative) etc. Q2.4 CMSC5711 Revision 1 V.x56.7b

As shown above, the image of an object has pixels of gray level with intensity from 0 to 5. Empty cells in the image represent that the pixel has intensity 0. A mean shift algorithm is applied to find the position of the object of size approximately 5x5 pixels. That means you need to find the centre of a 5x5 window best covering the object.  The initial window is centred at (x,y)=(4,6) as shown above. Find the location of the centre of the 5x5 search window at each step of the mean-shift algorithm. Show your calculation steps and round off numbers to integers during your calculations. What would happen if the initial search window is centred at (12,4). Q.2.5 1 2 3 4 5 6 7 8 9 10 11 12 13 14   CMSC5711 Revision 1 V.x56.7b

Q2.6 CMSC5711 Revision 1 V.x56.7b

Q2.6 (continue) CMSC5711 Revision 1 V.x56.7b