Problem Set 2 Reconstructing a Simpler World COS429 Computer Vision Due October (one week from today)13 th.

Slides:



Advertisements
Similar presentations
Y ( , ) R ( , ) T ( , ) -2 3 Y’ ( , ) R’ ( , ) T’ ( , )
Advertisements

COS 429 PS2: Reconstructing a Simpler World
Announcements. Structure-from-Motion Determining the 3-D structure of the world, and/or the motion of a camera using a sequence of images taken by a moving.
Hough Transforms CSE 6367 – Computer Vision Vassilis Athitsos University of Texas at Arlington.
CS 691 Computational Photography Instructor: Gianfranco Doretto 3D to 2D Projections.
Lecture 2: Convolution and edge detection CS4670: Computer Vision Noah Snavely From Sandlot ScienceSandlot Science.
Dynamic Occlusion Analysis in Optical Flow Fields
DREAM PLAN IDEA IMPLEMENTATION Introduction to Image Processing Dr. Kourosh Kiani
Paris town hall.
Mapping: Scaling Rotation Translation Warp
CS 4363/6353 INTRODUCTION TO COMPUTER GRAPHICS. WHAT YOU’LL SEE Interactive 3D computer graphics Real-time 2D, but mostly 3D OpenGL C/C++ (if you don’t.
1Ellen L. Walker Edges Humans easily understand “line drawings” as pictures.
Computer Vision REU Week 2 Adam Kavanaugh. Video Canny Put canny into a loop in order to process multiple frames of a video sequence Put canny into a.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Announcements Problem Set 2, handed out today, due next Tuesday. Late Homework should be turned into my office with date and time written on it. Mail problem.
Lecture 5: Projection CS6670: Computer Vision Noah Snavely.
CS485/685 Computer Vision Prof. George Bebis
Digital Image Processing
1 Image Filtering Readings: Ch 5: 5.4, 5.5, 5.6,5.7.3, 5.8 (This lecture does not follow the book.) Images by Pawan SinhaPawan Sinha formal terminology.
1 Lecture 12 Neighbourhood Operations (2) TK3813 DR MASRI AYOB.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
Lecture 2: Image filtering
17 VECTOR CALCULUS.
COS 429 PS3: Stitching a Panorama Due November 4 th.
Projectile Motion Physics 6A Prepared by Vince Zaccone
3D Scene Models Object recognition and scene understanding Krista Ehinger.
Projectile Motion Physics 6A Prepared by Vince Zaccone
October 8, 2013Computer Vision Lecture 11: The Hough Transform 1 Fitting Curve Models to Edges Most contours can be well described by combining several.
Neighborhood Operations
CS559: Computer Graphics Lecture 6: Edge Detection, Image Compositing, and Warping Li Zhang Spring 2010.
Week 2 - Wednesday CS361.
CS-424 Gregory Dudek Today’s Lecture Computational Vision –Images –Image formation in brief (+reading) –Image processing: filtering Linear filters Non-linear.
PARAMETRIC FUNCTIONS Today we will learn about parametric functions in the plane and analyze them using derivatives and integrals.
1-1 Measuring image motion velocity field “local” motion detectors only measure component of motion perpendicular to moving edge “aperture problem” 2D.
Feature-Based Parametric Modeling
Edge Detection Today’s reading Cipolla & Gee on edge detection (available online)Cipolla & Gee on edge detection From Sandlot ScienceSandlot Science.
CSC508 What You Should Be Doing Code, code, code –Programming Gaussian Convolution Sobel Edge Operator.
SIMD Image Processor Eric Liskay Andrew Northy Neraj Kumar 1.
CS654: Digital Image Analysis Lecture 25: Hough Transform Slide credits: Guillermo Sapiro, Mubarak Shah, Derek Hoiem.
Feature extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
CS 450: COMPUTER GRAPHICS PROJECTIONS SPRING 2015 DR. MICHAEL J. REALE.
Design Studies 20 ‘Show Off’ Project How to make a computer monitor In Google Sketchup By: Liam Jack.
CSC508 Convolution Operators. CSC508 Convolution Arguably the most fundamental operation of computer vision It’s a neighborhood operator –Similar to the.
Edge Detection and Geometric Primitive Extraction Jinxiang Chai.
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
Review on Graphics Basics. Outline Polygon rendering pipeline Affine transformations Projective transformations Lighting and shading From vertices to.
October 16, 2014Computer Vision Lecture 12: Image Segmentation II 1 Hough Transform The Hough transform is a very general technique for feature detection.
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
Features Jan-Michael Frahm.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
CS559: Computer Graphics Lecture 7: Image Warping and Morphing Li Zhang Spring 2010 Most slides borrowed from Yungyu ChuangYungyu Chuang.
Sliding Window Filters Longin Jan Latecki October 9, 2002.
CS 551 / 645: Introductory Computer Graphics Viewing Transforms.
COS 429 PS3: Stitching a Panorama Due November 10 th.
Terms and phrases you’ll need to analyze film!. Overview  Camera shots  Camera angles  Camera movement/Tracking  Editing  Sound.
3.7 Translations. A) Translation: when we SLIDE a figure to a different location. Transformation: an operation that maps or moves a figure onto an image.
Keypoint extraction: Corners 9300 Harris Corners Pkwy, Charlotte, NC.
Understanding Images. Pixels pixels Every image is made up of very small squares called pixels, and each pixel represents a color or shade. Pixels within.
Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.
Introduction to Computer Graphics
CMSC5711 Image processing and computer vision
PERSPECTIVE PROJECTION…...
CMSC5711 Revision 3 CMSC5711 revision 3 ver.x67.8c.
CMSC5711 Image processing and computer vision
Lecture 3: Camera Rotations and Homographies
First Homework One week
Lecture 2: Edge detection
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Problem Set 2 Reconstructing a Simpler World COS429 Computer Vision Due October (one week from today)13 th

Goal: Recover the 3D structure of the world

Problem 1: Making the World Simpler Simple World Assumptions: Flat surfaces that are either horizontal or vertical Objects rest on a white horizontal ground plane Task: Print Figure 1 and create objects for the world Take a picture of the world you created and add it to the report

Problem 2: Taking Orthographic Pictures Goal: Want pictures that preserve parallel lines from 3D to 2D Willing to accept weak perspective effects

Problem 2: Taking Orthographic Pictures Close up without zoom; closer edges are larger; Parallel lines in 3D are not parallel. Far away with zoom; edges are approximately equal in size and Parallel

Real life: OrthographicSelfie: Perspective

f f

Problem 2: Taking Orthographic Pictures Perspective: Take close-up pictures Orthographic: Take pictures from far away Use the zoom of the camera Crop the the central part of the image Requirement: Make the two images look as similar as possible. Submit the two pictures with the report

Problem 3: Orthographic Projection Goal: Projecting a 3D point to 2D (X, Y, Z) (x, y) Easy to do because we are reducing dimensions

X Y Z

X Y Z

X Y Z y x

X Y Z

X Y Z y x

X Y Z y x O(X = 0, Y = 0, Z = 0) o(x = 0, y = 0)

X Y Z y x O(X = 0, Y = 0, Z = 0) o(x = 0, y = 0)

x y Image Surface

x y P(x1, y1)

x y

x y

Problem 3: Orthographic Projection Projecting a 3D point to 2D (X, Y, Z) (x, y) Task: Prove the two projection equations below that relate the 3D world position (X, Y, Z) to the 2D projected camera position (x, y)

Problem 3: Orthographic Projection Projecting a 3D point to 2D (X, Y, Z) (x, y) Task: Prove the two projection equations below that relate the 3D world position (X, Y, Z) to the 2D projected camera position (x, y) X + x 0 cos(  ) Y – sin(  ) Z + y 0 x = y = image coordinates World coordinates

Problem 3: Orthographic Projection Projecting a 3D point to 2D (X, Y, Z) (x, y) Task: Prove the two projection equations below that relate the 3D world position (X, Y, Z) to the 2D projected camera position (x, y) Dimension Reduction Rotation 2D Translation

Problem 4, 5, 6: Recover 3D points from 2D

Problem 4: Edge Constraints Crack the problem by first looking at special points with special features

Problem 4: Edge Constraints Extract edges from the image: I(x, y) =

Problem 4: Edge Constraints How to identify an edge? Edge in 2D

Problem 4: Edge Constraints How to identify an edge? “Edge” in 1D

The Gradient

The Sobel operator

Using the The Sobel operator to compute derivatives (and detect edges)

Problem 4: Edge Constraints Edge strength: Edge Orientation:

Problem 4: Edge Constraints

Vertical Edges in 3D?

Problem 4: Edge Constraints Horizontal Edges in 3D?

Problem 4: Edge Constraints

For 2D points on vertical edges:

Problem 4: Edge Constraints For 2D points on vertical edges: For 2D points on horizontal edges:

Problem 4: Edge Constraints For 2D points on vertical edges: For 2D points on horizontal edges:

Problem 4: Edge Constraints Constraints for points on edges?

Problem 4: Edge Constraints Constraints for points on edges? Z(x,y) for points on the vertical edges? Z(x,y) for points on the horizontal edges?

Problem 4: Ground Constraints Color threshold determines ground from objects On the ground Y(x, y) = 0

Problem 5: Surface Constraints Constraints for points on flat surfaces: Approximating second derivatives:

Problem 5: Fill Missing Code for Constraints Want to use constraints from Problem 4 and 5 to determine Y(x, y) and Z(x, y) Two constraints missing from existing code Task: Write two lines of code (lines 171 and 187) Copy these two lines and add them to the report

Problem 6: A Simple Inference Scheme Write the constraints as a system of linear equations Task: Run simpleworldY.m to generate images for the report Include some screen shots of the generated figures and include in report

Extra Credit 1: Violating Simple World Assumptions What if we violate our assumptions? Show examples where the reconstruction fails Why does it fail?

Extra Credit 2: The Real World Take pictures of the real world How can we modify this assignment to getter better 3D reconstruction in the real world? Try to handle a few more situations Possible final project?

What to Submit: One PDF file report One ZIP file containing all the source code, and a “simpleworldY.m” file that takes no parameters as input and runs directly in Matlab to generate the results reported in your PDF file.

PDF Report (1) Take a picture of the world you created (2) Submit two pictures – one showing orthographic projection and the other perspective projection (3) Prove the two projection equations (4) Write the constraints for Z(x, y) (5) Fill in missing kernels (lines 171 and 187) and copy code into report (6) Show results and figures output by simpleworldY.m [Optional] Extra credit