Chapter 13 Image-Based Rendering

Slides:



Advertisements
Similar presentations
Parameterized Environment Maps
Advertisements

Lecture 8 Transparency, Mirroring
Week 10 - Monday.  What did we talk about last time?  Global illumination  Shadows  Projection shadows  Soft shadows.
Acquiring the Reflectance Field of a Human Face Paul Debevec, Tim Hawkins, Chris Tchou, Haarm-Pieter Duiker, Westley Sarokin, Mark Sagar Haarm-Pieter Duiker,
Image Compositing and Matting. Introduction Matting and compositing are important operations in the production of special effects. These techniques enable.
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
18.1 Si31_2001 SI31 Advanced Computer Graphics AGR Lecture 18 Image-based Rendering Light Maps What We Did Not Cover Learning More...
Advanced Computer Graphics (Spring 2005) COMS 4162, Lecture 21: Image-Based Rendering Ravi Ramamoorthi
Measurement, Inverse Rendering COMS , Lecture 4.
Image-Based Rendering Computer Vision CSE576, Spring 2005 Richard Szeliski.
High-Quality Video View Interpolation
Copyright  Philipp Slusallek IBR: View Interpolation Philipp Slusallek.
CSCE 641 Computer Graphics: Image-based Rendering (cont.) Jinxiang Chai.
1 Lecture 11 Scene Modeling. 2 Multiple Orthographic Views The easiest way is to project the scene with parallel orthographic projections. Fast rendering.
Computational Photography Light Field Rendering Jinxiang Chai.
CSE473/573 – Stereo Correspondence
CSCE 641: Computer Graphics Image-based Rendering Jinxiang Chai.
CS 563 Advanced Topics in Computer Graphics View Interpolation and Image Warping by Brad Goodwin Images in this presentation are used WITHOUT permission.
The Story So Far The algorithms presented so far exploit: –Sparse sets of images (some data may not be available) –User help with correspondences (time.
Computer Graphics Inf4/MSc Computer Graphics Lecture 11 Texture Mapping.
The Uncanny Valley in 3D Modeling Steve Seitz and Rick Szeliski BIRS Workshop on Computer Vision and the Internet August 31, 2009.
1 Computer Graphics Week13 –Shading Models. Shading Models Flat Shading Model: In this technique, each surface is assumed to have one normal vector (usually.
What Does the Scene Look Like From a Scene Point? Donald Tanguay August 7, 2002 M. Irani, T. Hassner, and P. Anandan ECCV 2002.
Lecture 11 Stereo Reconstruction I Lecture 11 Stereo Reconstruction I Mata kuliah: T Computer Vision Tahun: 2010.
09/09/03CS679 - Fall Copyright Univ. of Wisconsin Last Time Event management Lag Group assignment has happened, like it or not.
Image-based rendering Michael F. Cohen Microsoft Research.
University of Palestine Faculty of Applied Engineering and Urban Planning Software Engineering Department Introduction to computer vision Chapter 2: Image.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
Image-based Rendering. © 2002 James K. Hahn2 Image-based Rendering Usually based on 2-D imagesUsually based on 2-D images Pre-calculationPre-calculation.
03/24/03© 2003 University of Wisconsin Last Time Image Based Rendering from Sparse Data.
CS-378: Game Technology Lecture #4: Texture and Other Maps Prof. Okan Arikan University of Texas, Austin V Lecture #4: Texture and Other Maps.
Image Based Rendering. Light Field Gershun in 1936 –An illuminated objects fills the surrounding space with light reflected of its surface, establishing.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
112/5/ :54 Graphics II Image Based Rendering Session 11.
Computer vision: models, learning and inference M Ahad Multiple Cameras
CSCE 641 Computer Graphics: Image-based Rendering (cont.) Jinxiang Chai.
1Ellen L. Walker 3D Vision Why? The world is 3D Not all useful information is readily available in 2D Why so hard? “Inverse problem”: one image = many.
1 CSCE 441: Computer Graphics Hidden Surface Removal Jinxiang Chai.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Eigen Texture Method : Appearance compression based method Surface Light Fields for 3D photography Presented by Youngihn Kho.
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Computer vision: geometric models Md. Atiqur Rahman Ahad Based on: Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Presented by 翁丞世  View Interpolation  Layered Depth Images  Light Fields and Lumigraphs  Environment Mattes  Video-Based.
Advanced Computer Graphics
Real-time Walkthrough of Virtual Space using Environment Map
Computer Graphics Chapter 9 Rendering.
Jun Shimamura, Naokazu Yokoya, Haruo Takemura and Kazumasa Yamazawa
Steps Towards the Convergence of Graphics, Vision, and Video
Image-Based Rendering
Image-Based Rendering
3D Graphics Rendering PPT By Ricardo Veguilla.
Removing Highlight Spots in Visual Hull Rendering
Chapter 10: Computer Graphics
Interactive Computer Graphics
Models and Architectures
© University of Wisconsin, CS559 Fall 2004
© 2005 University of Wisconsin
Three-Dimensional Concepts. Three Dimensional Graphics  It is the field of computer graphics that deals with generating and displaying three dimensional.
Models and Architectures
CSCE 441: Computer Graphics Hidden Surface Removal
Models and Architectures
Introduction to Computer Graphics with WebGL
Common Classification Tasks
Image Based Modeling and Rendering (PI: Malik)
(c) 2002 University of Wisconsin
UMBC Graphics for Games
Models and Architectures
CS-378: Game Technology Lecture #4: Texture and Other Maps
Models and Architectures
Presentation transcript:

Chapter 13 Image-Based Rendering Guan-Shiuan Kuo (郭冠軒) 0988039217 r07942063@ntu.edu.tw

Outline 13.1 View Interpolation 13.2 Layered Depth Images 13.3 Light Fields and Lumigraphs 13.4 Environment Mattes 13.5 Video-Based Rendering

13.1 View Interpolation 13.1.1 View-Dependent Texture Maps 13.1.2 Application: Photo Tourism

Intel® freeD Technology https://www.youtube.com/watch?v=J7xIBoPr83A

13.1 View Interpolation View interpolation creates a seamless transition between a pair of reference images using one or more pre-computed depth maps. Closely related to this idea are view-dependent texture maps, which blend multiple texture maps on a 3D model’s surface.

13.1 View Interpolation

13.1.1 View-Dependent Texture Maps View-dependent texture maps are closely related to view interpolation. Instead of associating a separate depth map with each input image, a single 3D model is created for the scene, but different images are used as texture map sources depending on the virtual camera’s current position.

13.1.1 View-Dependent Texture Maps

13.1.1 View-Dependent Texture Maps Given a new virtual camera position, the similarity of this camera’s view of each polygon (or pixel) is compared with that of potential source images. Even though the geometric model can be fairly coarse, blending between different views gives a strong sense of more detailed geometry because of the parallax (visual motion) between corresponding pixels.

13.1.2 Application: Photo Tourism

13.1.2 Application: Photo Tourism https://www.youtube.com/watch?v=mTBPGuPLI5Y

13.2 Layered Depth Images 13.2.1 Imposters, Sprites, and Layers

13.2 Layered Depth Images When a depth map is warped to a novel view, holes and cracks inevitably appear behind the foreground objects. One way to alleviate this problem is to keep several depth and color values (depth pixels) at every pixel in a reference image (or, at least for pixels near foreground–background transitions).

13.2 Layered Depth Images The resulting data structure, which is called a Layered Depth Image (LDI), can be used to render new views using a back-to-front forward warping algorithm (Shade, Gortler, He et al. 1998).

13.2 Layered Depth Images

13.2.1 Imposters, Sprites, and Layers An alternative to keeping lists of color-depth values at each pixel, as is done in the LDI, is to organize objects into different layers or sprites.

13.3 Light Fields and Lumigraphs 13.3.1 Unstructured Lumigraph 13.3.2 Surface Light Fields

13.3 Light Fields and Lumigraphs Is it possible to capture and render the appearance of a scene from all possible viewpoints and, if so, what is the complexity of the resulting structure? Let us assume that we are looking at a static scene, i.e., one where the objects and illuminants are fixed, and only the observer is moving around.

13.3 Light Fields and Lumigraphs We can describe each image by the location and orientation of the virtual camera (6 DOF). DOF: degree of freedom ( 𝑡 𝑥 , 𝑡 𝑦 , 𝑡 𝑧 , 𝑟 𝑥 , 𝑟 𝑦 , 𝑟 𝑧 ) If we capture a two-dimensional spherical image around each possible camera location, we can re- render any view from this information.

13.3 Light Fields and Lumigraphs To make the parameterization of this 4D function simpler, let us put two planes in the 3D scene roughly bounding the area of interest. Any light ray terminating at a camera that lives in front of the s-t plane (assuming that this space is empty) passes through the two planes at (𝑠,𝑡) and (𝑢,𝑣) and can be described by its 4D coordinate (𝑠,𝑡,𝑢,𝑣).

13.3 Light Fields and Lumigraphs

13.3 Light Fields and Lumigraphs This diagram (and parameterization) can be interpreted as describing a family of cameras living on the 𝑠-𝑡 plane with their image planes being the 𝑢-𝑣 plane. The 𝑢-𝑣 plane can be placed at infinity, which corresponds to all the virtual cameras looking in the same direction.

13.3 Light Fields and Lumigraphs While a light field can be used to render a complex 3D scene from novel viewpoints, a much better rendering (with less ghosting) can be obtained if something is known about its 3D geometry. The Lumigraph system of Gortler, Grzeszczuk, Szeliski et al. (1996) extends the basic light field rendering approach by taking into account the 3D location of surface points corresponding to each 3D ray.

13.3 Light Fields and Lumigraphs Consider the ray (𝑠,𝑢) corresponding to the dashed line in Figure 13.8, which intersects the object’s surface at a distance 𝑧 from the 𝑢-𝑣 plane. Instead of using quadri-linear interpolation of the nearest sampled (𝑠,𝑡,𝑢,𝑣) values around a given ray to determine its color, the (𝑢,𝑣) values are modified for each discrete ( 𝑠 𝑖 , 𝑡 𝑖 ) camera.

13.3 Light Fields and Lumigraphs

13.3.1 Unstructured Lumigraph When the images in a Lumigraph are acquired in an unstructured (irregular) manner, it can be counterproductive to resample the resulting light rays into a regularly binned (𝑠,𝑡,𝑢,𝑣) data structure. This is both because resampling always introduces a certain amount of aliasing and because the resulting gridded light field can be populated very sparsely or irregularly.

13.3.1 Unstructured Lumigraph The alternative is to render directly from the acquired images, by finding for each light ray in a virtual camera the closest pixels in the original images.

13.3.1 Unstructured Lumigraph The unstructured Lumigraph rendering (ULR) system of describes how to select such pixels by combining a number of fidelity criteria. epipole consistency (distance of rays to a source camera’s center) angular deviation (similar incidence direction on the surface) resolution (similar sampling density along the surface) continuity (to nearby pixels) consistency (along the ray)

13.3.2 Surface Light Fields If we know the 3D shape of the object or scene whose light field is being modeled, we can effectively compress the field because nearby rays emanating from nearby surface elements have similar color values. Nearby Lumispheres will be highly correlated and hence amenable to both compression and manipulation.

13.3.2 Surface Light Fields

13.3.2 Surface Light Fields To estimate the diffuse component of each Lumisphere, a median filtering over all visible exiting directions is first performed for each channel. Once this has been subtracted from the Lumisphere, the remaining values, which should consist mostly of the specular components, are reflected around the local surface normal, which turns each Lumisphere into a copy of the local environment around that point.

13.4 Environment Mattes 13.4.1 Higher-Dimensional Light Fields 13.4.2 The Modeling to Rendering Continuum

13.4 Environment Mattes What if instead of moving around a virtual camera, we take a complex, refractive object, such as the water goblet shown in Figure 13.10, and place it in front of a new background? Instead of modeling the 4D space of rays emanating from a scene, we now need to model how each pixel in our view of this object refracts incident light coming from its environment.

13.4 Environment Mattes

13.4 Environment Mattes If we assume that other objects and illuminants are sufficiently distant (the same assumption we made for surface light fields, 4D mapping (𝑥,𝑦)→ 𝜙,𝜃 captures all the information between a refractive object and its environment.

13.4 Environment Mattes Zongker, Werner, Curless et al. (1999) call such a representation an environment matte, since it generalizes the process of object matting to not only cut and paste an object from one image into another but also take into account the subtle refractive or reflective interplay between the object and its environment.

13.4.1 Higher-Dimensional Light Fields An environment matte in principle maps every pixel (𝑥,𝑦) into a 4D distribution over light rays and is, hence, a six-dimensional representation. If we want to handle dynamic light fields, we need to add another temporal dimension. Similarly, if we want a continuous distribution over wavelengths, this becomes another dimension.

13.4.2 The Modeling to Rendering Continuum The image-based rendering representations and algorithms we have studied in this chapter span a continuum ranging from classic 3D texture-mapped models all the way to pure sampled ray-based representations such as light fields.

13.4.2 The Modeling to Rendering Continuum

13.4.2 The Modeling to Rendering Continuum Representations such as view-dependent texture maps and Lumigraphs still use a single global geometric model, but select the colors to map onto these surfaces from nearby images. The best choice of representation and rendering algorithm depends on both the quantity and quality of the input imagery as well as the intended application.

13.5 Video-Based Rendering 13.5.1 Video-Based Animation 13.5.2 Video Textures 13.5.3 Application: Animating Pictures 13.5.4 3D Video 13.5.5 Application: Video-Based Walkthroughs

13.5 Video-Based Rendering A fair amount of work has been done in the area of video-based rendering and video-based animation, two terms first introduced by Schödl, Szeliski, Salesin et al. (2000) to denote the process of generating new video sequences from captured video footage.

13.5 Video-Based Rendering We start with video-based animation (Section 13.5.1), in which video footage is re-arranged or modified, e.g., in the capture and re-rendering of facial expressions. Next, we turn our attention to 3D video (Section 13.5.4), in which multiple synchronized video cameras are used to film a scene from different directions.

13.5.1 Video-Based Animation An early example of video-based animation is Video Rewrite, in which frames from original video footage are rearranged in order to match them to novel spoken utterances, e.g., for movie dubbing (Figure 13.12). This is similar in spirit to the way that concatenative speech synthesis systems work (Taylor 2009).

13.5.1 Video-Based Animation

13.5.2 Video Textures Video texture is a short video clip that can be arbitrarily extended by re-arranging video frames while preserving visual continuity (Schödl, Szeliski, Salesin et al. 2000). The simplest approach is to match frames by visual similarity and to jump between frames that appear similar.

13.5.3 Application: Animating Pictures https://www.youtube.com/watch?v=ZVrYyX3bHI8

13.5.4 3D Video In recent years, the popularity of 3D movies has grown dramatically, with recent releases ranging from Hannah Montana, through U2’s 3D concert movie, to James Cameron’s Avatar. Currently, such releases are filmed using stereoscopic camera rigs and displayed in theaters (or at home) to viewers wearing polarized glasses.

13.5.4 3D Video The stereo matching techniques developed in the computer vision community along with image- based rendering (view interpolation) techniques from graphics are both essential components in such scenarios, which are sometimes called free- viewpoint video (Carranza, Theobalt, Magnor et al. 2003) or virtual viewpoint video (Zitnick, Kang, Uyttendaele et al. 2004).

13.5.4 3D Video

13.5.5 Application: Video-Based Walkthroughs https://www.youtube.com/watch?v=8EfCf1Xt5yA