Automated 3D Model Construction for Urban Environments Christian Frueh John Flynn Avideh Zakhor Next Generation 4D Distributed Modeling and Visualization.

Slides:



Advertisements
Similar presentations
REQUIRING A SPATIAL REFERENCE THE: NEED FOR RECTIFICATION.
Advertisements

Feature Based Image Mosaicing
The fundamental matrix F
CSE473/573 – Stereo and Multiple View Geometry
Efficient High-Resolution Stereo Matching using Local Plane Sweeps Sudipta N. Sinha, Daniel Scharstein, Richard CVPR 2014 Yongho Shin.
For Internal Use Only. © CT T IN EM. All rights reserved. 3D Reconstruction Using Aerial Images A Dense Structure from Motion pipeline Ramakrishna Vedantam.
www-video.eecs.berkeley.edu/research
3D Graphics Rendering and Terrain Modeling
Automatic Feature Extraction for Multi-view 3D Face Recognition
Intelligent Systems Lab. Extrinsic Self Calibration of a Camera and a 3D Laser Range Finder from Natural Scenes Davide Scaramuzza, Ahad Harati, and Roland.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Semi-automatic Range to Range Registration: A Feature-based Method Chao Chen & Ioannis Stamos Computer Science Department Graduate Center, Hunter College.
Localization of Piled Boxes by Means of the Hough Transform Dimitrios Katsoulas Institute for Pattern Recognition and Image Processing University of Freiburg.
Shape from Contours and Multiple Stereo A Hierarchical, Mesh-Based Approach Hendrik Kück, Wolfgang Heidrich, Christian Vogelgsang.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
Last Time Pinhole camera model, projection
Copyright  Philipp Slusallek Cs fall IBR: Model-based Methods Philipp Slusallek.
1 MURI review meeting 09/21/2004 Dynamic Scene Modeling Video and Image Processing Lab University of California, Berkeley Christian Frueh Avideh Zakhor.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2005 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Contents Description of the big picture Theoretical background on this work The Algorithm Examples.
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
High-Quality Video View Interpolation
Lecture 11: Structure from motion CS6670: Computer Vision Noah Snavely.
The plan for today Camera matrix
Stereo and Multiview Sequence Processing. Outline Stereopsis Stereo Imaging Principle Disparity Estimation Intermediate View Synthesis Stereo Sequence.
Multiple View Geometry : Computational Photography Alexei Efros, CMU, Fall 2006 © Martin Quinn …with a lot of slides stolen from Steve Seitz and.
Introduction to 3D Graphics John E. Laird. Basic Issues u Given a internal model of a 3D world, with textures and light sources how do you project it.
3-D Scene u u’u’ Study the mathematical relations between corresponding image points. “Corresponding” means originated from the same 3D point. Objective.
Computer Vision Spring ,-685 Instructor: S. Narasimhan WH 5409 T-R 10:30am – 11:50am Lecture #15.
Yuping Lin and Gérard Medioni.  Introduction  Method  Register UAV streams to a global reference image ▪ Consecutive UAV image registration ▪ UAV to.
Technology and Historical Overview. Introduction to 3d Computer Graphics  3D computer graphics is the science, study, and method of projecting a mathematical.
What we didn’t have time for CS664 Lecture 26 Thursday 12/02/04 Some slides c/o Dan Huttenlocher, Stefano Soatto, Sebastian Thrun.
Shape-Based Human Detection and Segmentation via Hierarchical Part- Template Matching Zhe Lin, Member, IEEE Larry S. Davis, Fellow, IEEE IEEE TRANSACTIONS.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
1. Introduction Motion Segmentation The Affine Motion Model Contour Extraction & Shape Estimation Recursive Shape Estimation & Motion Estimation Occlusion.
3D SLAM for Omni-directional Camera
MESA LAB Multi-view image stitching Guimei Zhang MESA LAB MESA (Mechatronics, Embedded Systems and Automation) LAB School of Engineering, University of.
High-Resolution Interactive Panoramas with MPEG-4 발표자 : 김영백 임베디드시스템연구실.
CS654: Digital Image Analysis Lecture 8: Stereo Imaging.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
CS 4487/6587 Algorithms for Image Analysis
Summary of “Data Processing Algorithm for Generating 3D building Facade Meshes From Laser Scans and Camera Images”. An article by Christian Frueh, Siddharth.
Spring 2015 CSc 83020: 3D Photography Prof. Ioannis Stamos Mondays 4:15 – 6:15
Vehicle Segmentation and Tracking From a Low-Angle Off-Axis Camera Neeraj K. Kanhere Committee members Dr. Stanley Birchfield Dr. Robert Schalkoff Dr.
Introduction to Soft Copy Photogrammetry
Raquel A. Romano 1 Scientific Computing Seminar May 12, 2004 Projective Geometry for Computer Vision Projective Geometry for Computer Vision Raquel A.
Chapter 5 Multi-Cue 3D Model- Based Object Tracking Geoffrey Taylor Lindsay Kleeman Intelligent Robotics Research Centre (IRRC) Department of Electrical.
112/5/ :54 Graphics II Image Based Rendering Session 11.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Yizhou Yu Texture-Mapping Real Scenes from Photographs Yizhou Yu Computer Science Division University of California at Berkeley Yizhou Yu Computer Science.
High Resolution Surface Reconstruction from Overlapping Multiple-Views
Presented by: Idan Aharoni
John Morris Stereo Vision (continued) Iolanthe returns to the Waitemata Harbour.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Model Refinement from Planar Parallax Anthony DickRoberto Cipolla Department of Engineering University of Cambridge.
Computer Vision Computer Vision based Hole Filling Chad Hantak COMP December 9, 2003.
Integrating LiDAR Intensity and Elevation Data for Terrain Characterization in a Forested Area Cheng Wang and Nancy F. Glenn IEEE GEOSCIENCE AND REMOTE.
Acquiring, Stitching and Blending Diffuse Appearance Attributes on 3D Models C. Rocchini, P. Cignoni, C. Montani, R. Scopigno Istituto Scienza e Tecnologia.
A Plane-Based Approach to Mondrian Stereo Matching
Paper – Stephen Se, David Lowe, Jim Little
3D Graphics Rendering PPT By Ricardo Veguilla.
The Brightness Constraint
Common Classification Tasks
Vehicle Segmentation and Tracking in the Presence of Occlusions
The Brightness Constraint
Image Based Modeling and Rendering (PI: Malik)
Noah Snavely.
Filtering An image as a function Digital vs. continuous images
Presentation transcript:

Automated 3D Model Construction for Urban Environments Christian Frueh John Flynn Avideh Zakhor Next Generation 4D Distributed Modeling and Visualization University of California, Berkeley June 13, 2002

Presentation Overview  Introduction  Ground based modeling - Mesh processing  Airborne modeling - Aerial photos - Airborne laser scans  3D Model Fusion  Rendering  Conclusion and Future Work

Introduction Goal: Generate 3D model of a city for virtual walk/drive/fly-thrus and simulations Fast Automated Photorealistic 3D model of street scenery & building façades highly detailed needed: 3D Model of terrain and buildings tops & sides coarse resolution For Fly-Thru: For Walk/Drive-Thru:

Introduction 3D model of building façades 3D Model of terrain and building tops Airborne Modeling Laser scans/images from plane Ground Based Modeling Laser scans & images from acquisition vehicle Complete 3D City Model Fusion

Airborne Modeling Available Data: Aerial Photos Airborne laser scans Acquisition of terrain shape and top-view building geometry Goal: Geometry: 2 approaches: I) stereo matching of photos Texture: from aerial photos II) airborne laser scans

Airborne Modeling Stereo photo pairs from city/urban areas, ~ 60% overlap Approach I : Stereo Matching Camera parameter computation, Matching, Distortion reduction, Model generation Segmentation Manual: Automated: Semi-Automatic (last year)

Stereo Matching Stereo pair from downtown Berkeley and the estimated disparity after removing perspective distortions

Stereo Matching Results Downtown Oakland

Airborne Modeling Scanning city from plane point cloud Resolution 1 scan point/m 2 Berkeley: 40 Million scan points Approach II: Airborne Laser Scans

Airborne Laser Scans Re-sampling point cloud Sorting into grid Filling holes usable for: Map-like height field Mesh Generation Monte Carlo Localization

Textured Mesh Generation 1. Connecting grid vertices to mesh 2. Applying Q-slim simplification 3. Texture mapping: Semi-automatic Manual selection of few correspondence points: 10 mins/entire Berkeley Automated camera pose estimation Automated computation of texture for mesh

Airborne Model East Berkeley campus with campanile

Airborne Model Downtown Berkeley

Ground Based Modeling buildings truck x z y 2D laser v u  Acquisition vehicle Truck with rack: 2 fast 2D laser scanners digital camera Acquisition of highly detailed 3D building façade models Scanning setup vertical 2D laser scanner for geometry capture horizontal scanner for pose estimation Goal:

Scan Matching & Initial Path Computation Horizontal laser scans: Continuously captured during vehicle motion t = t 0 t = t 1 Overlap Relative position estimation by scan-to-scan matching Scan matching Translation (  u,  v) Rotation  (  u,  v) Adding relative steps (  u i,  v i,  i ) path (x i,y i,  i ) (  u 1,  v 1,  1 ) (  u 2,  v 2,  2 ) … (  u i,  v i,  i ) (  u i-1,  v i-1,  i-1 )  3 DOF pose (x, y, yaw)

6 DOF Pose Estimation From Images  Scan matching cannot estimate vertical motion - Small bumps and rolls - Slopes in hill areas  Full 6 DOF pose of the vehicle is important; affects: - Future processing of the 3D and intensity data - Texture mapping of the resulting 3D models  Extend initial 3 DOF pose by deriving missing 3 DOF (z, pitch, roll) from images

6 DOF Pose Estimation From Images Central idea: photo-consistency Each 3D scan point can be projected into images using initial 3 DOF pose If pose estimate is correct, point should appear the same in all images Use discrepancies in projected position of 3D points within multiple images to solve for the full pose

6 DOF Pose Estimation – Algorithm  3DOF of laser as initial estimate  Project scan points into both images  If not consistent, use image correlation to find correct projection  Ransac used for robustness

6 DOF Pose Estimation – Algorithm  Use 3 DOF pose of laser scan matching as initial estimate  Some scan points are taken simultaneously with each image, the exact projection of these scan points are know in this image  Image correlation to find the projection of these scan points in other images - Not always correct results - Many outliers  Reprojected error is minimized across many images to find the pose of each image  RANSAC for robustness

6 DOF Pose Estimation – Results with 3 DOF pose with 6 DOF pose

6 DOF Pose Estimation – Results

Monte Carlo Localization (1) Previously: Global 3 DOF pose correction using aerial photography b) path after MCL correction a) path before MCL correction After correction, points fit to edges of aerial image

Monte Carlo Localization (2) Extend MCL to work with airborne laser data and 6 DOF pose No perspective shifts of building tops, no shadow lines Use terrain shape to estimate z coordinate of truck Fewer particles necessary, increased computation speed Significantly higher accuracy near high buildings and tree areas Correct additional DOF for vehicle pose (z, pitch, roll) Modeling not restricted to flat areas Now:

Monte Carlo Localization (3) Resulting corrected path overlaid with airborne laser height field Track global 3D position of vehicle to correct relative 6 DOF motion estimates

Path Segmentation vertical scans: 107,082 scan points: ~ 15 million 24 mins, 6769 meters Too large to process as one block! Segment path into quasi- linear pieces Cut path at curves and empty areas Remove redundant segments

Path Segmentation Resulting path segments overlaid with edges of airborne laser height map

Simple Mesh Generation

Triangulate Point cloud Mesh Problem: Partially captured foreground objects erroneous scan points due to glass reflection Side views look “noisy”Remove foreground: extract facades

Façade Extraction and Processing (1) 1. Transform path segment into depth image 2. Histogram analysis over vertical scans scan nr depth split depth depth local minimum main depth scanner split depth ground points depth value s n,υ for a scan point P n,υ main depth

Façade Extraction and Processing (2) 3. Separate depth image into 2 layers: background =building facades foreground =trees, cars etc.

Façade Extraction and Processing (3) 4. Process background layer: Fill areas occluded by foreground objects by extending geometry from boundaries Horizontal, vertical, planar interpolation, RANSAC Detect and remove invalid scan points Apply segmentation Fill remaining holes in large segments Final result: “clean” background layer Remove isolated segments

Façade Extraction – Examples (1) without processing with processing

Façade Extraction – Examples (2) with processing without processing

Façade Extraction – Examples (3) without processing with processing

Facade Processing

Foreground Removal

Mesh Generation Downtown Berkeley

Automatic Texture Mapping (1) Camera calibrated and synchronized with laser scanners Transformation matrix between camera image and laser scan vertices can be computed 1. Project geometry into images 2. Mark occluding foreground objects in image 3. For each background triangle: Search pictures in which triangle is not occluded, and texture with corresponding picture area

Automatic Texture Mapping(2) Efficient representation: texture atlas Copy texture of all triangles into “mosaic” image Typical texture reduction: factor 8..12

Automatic Texture Mapping (3) Large foreground objects: Some of the filled-in triangles are not visible in any image! “texture holes” in the atlas Texture synthesis: preliminary Mark holes corresponding to non- textured triangles in the atlas Search the image for areas matching the hole boundaries Fill the hole by copying missing pixels from these image

Automatic Texture Mapping (4) Texture holes marked Texture holes filled

Automatic Texture Mapping(5)

Ground Based Modeling - Results Façade models of downtown Berkeley

Ground Based Modeling - Results Façade models of downtown Berkeley

Model Fusion Goal: Fusion of ground based and airborne model to one single model 1.Registration of models 2.Combining the registered meshes Model Fusion: Façade model Airborne model

Registration of Models Models are already registered with each via Monte-Carlo-Localization ! Which model to use where?

Preparing Ground Based Models Intersect path segments with each other Remove degenerated, redundant triangles in overlapping areas original mesh redundant triangles removed

Preparing Airborne Model  Remove facades in airborne model where ground based geometry is available  Add ground based façades  Fill remaining gaps with a “blend mesh” to hide model transitions Ground based model has 5-10 times higher resolution

Preparing Airborne Model Initial airborne model

Preparing Airborne Model Remove facades where ground based geometry is available

Combining Models Add ground based façade models

Combining Models Fill remaining gaps with a “blend mesh” to hide model transitions

Model Fusion - Results

Rendering  Subdivide model and create multiple level-of-details (LOD)  Generate scene graph, decide which LOD to render when Up to 270,000 triangles, 20 MB texture per path segment 4.28 million triangles, 348 MB texture for 4 downtown blocks Ground based models: Difficult to render interactively!

Multiple LODs for façade meshes Highest LOD Geometry: 10% Texture: 25% Lower LOD of original mesh Qslim mesh simplification Texture subsampling

Façade Model Subdivision for Rendering Subdivide 2 highest LODs of façade meshes along cut planes Sub- scene LOD 2 Path segment Global scene LOD 0 LOD 1 LOD 0 LOD 1 LOD 0 LOD 1 Submesh

Interactive Rendering Downtown blocks: Interactive rendering with web-based browser!

Future Work  Resolution enhancement and post- processing of LIDAR data  Devise new data acquisition system and algorithms to capture - both sides of street simultaneously - texture for upper parts of tall buildings  Include foreground objects in model  Add temporal component to dynamically update models  Compact representation  Interactive rendering