Introduction Autostereoscopic displays give great promise as the future of 3D technology. These displays spatially multiplex many views onto a screen,

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

S INGLE -I MAGE R EFOCUSING AND D EFOCUSING Wei Zhang, Nember, IEEE, and Wai-Kuen Cham, Senior Member, IEEE.
Computational Biology, Part 23 Biological Imaging II Robert F. Murphy Copyright  1996, 1999, All rights reserved.
Leveraging Stereopsis for Saliency Analysis
Hongliang Li, Senior Member, IEEE, Linfeng Xu, Member, IEEE, and Guanghui Liu Face Hallucination via Similarity Constraints.
嵌入式視覺 Feature Extraction
Document Image Processing
CHAPTER 12 Height Maps, Hidden Surface Removal, Clipping and Level of Detail Algorithms © 2008 Cengage Learning EMEA.
CMPE 466 COMPUTER GRAPHICS Chapter 8 2D Viewing Instructor: D. Arifler Material based on - Computer Graphics with OpenGL ®, Fourth Edition by Donald Hearn,
Video Inpainting Under Constrained Camera Motion Kedar A. Patwardhan, Student Member, IEEE, Guillermo Sapiro, Senior Member, IEEE, and Marcelo Bertalm.
September 10, 2013Computer Vision Lecture 3: Binary Image Processing 1Thresholding Here, the right image is created from the left image by thresholding,
Image Indexing and Retrieval using Moment Invariants Imran Ahmad School of Computer Science University of Windsor – Canada.
Ai-Mei Huang and Truong Nguyen Video Processing LabECE Dept, UCSD, La Jolla, CA This paper appears in: Image Processing, ICIP IEEE International.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
Segmentation Divide the image into segments. Each segment:
Detecting Image Region Duplication Using SIFT Features March 16, ICASSP 2010 Dallas, TX Xunyu Pan and Siwei Lyu Computer Science Department University.
Virtual Control of Optical Axis of the 3DTV Camera for Reducing Visual Fatigue in Stereoscopic 3DTV Presenter: Yi Shi & Saul Rodriguez March 26, 2008.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Computer Vision Lecture 3: Digital Images
Fast Texture Synthesis Tree-structure Vector Quantization Li-Yi WeiMarc Levoy Stanford University.
E.G.M. PetrakisBinary Image Processing1 Binary Image Analysis Segmentation produces homogenous regions –each region has uniform gray-level –each region.
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
Image Analogies Aaron Hertzmann (1,2) Charles E. Jacobs (2) Nuria Oliver (2) Brian Curless (3) David H. Salesin (2,3) 1 New York University 1 New York.
 C. C. Hung, H. Ijaz, E. Jung, and B.-C. Kuo # School of Computing and Software Engineering Southern Polytechnic State University, Marietta, Georgia USA.
Tricolor Attenuation Model for Shadow Detection. INTRODUCTION Shadows may cause some undesirable problems in many computer vision and image analysis tasks,
CS 376b Introduction to Computer Vision 04 / 29 / 2008 Instructor: Michael Eckmann.
New Segmentation Methods Advisor : 丁建均 Jian-Jiun Ding Presenter : 蔡佳豪 Chia-Hao Tsai Date: Digital Image and Signal Processing Lab Graduate Institute.
ICPR/WDIA-2012 High Quality Novel View Synthesis Based on Low Resolution Depth Image and High Resolution Color Image Jui-Chiu Chiang, Zheng-Feng Liu, and.
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
Boris Babenko Department of Computer Science and Engineering University of California, San Diego Semi-supervised and Unsupervised Feature Scaling.
Advanced Computer Technology II FTV and 3DV KyungHee Univ. Master Course Kim Kyung Yong 10/10/2015.
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
CS 376b Introduction to Computer Vision 02 / 22 / 2008 Instructor: Michael Eckmann.
Under Supervision of Dr. Kamel A. Arram Eng. Lamiaa Said Wed
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Digital Image Processing CCS331 Relationships of Pixel 1.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Ray Divergence-Based Bundle Adjustment Conditioning for Multi-View Stereo Mauricio Hess-Flores 1, Daniel Knoblauch 2, Mark A. Duchaineau 3, Kenneth I.
A genetic approach to the automatic clustering problem Author : Lin Yu Tseng Shiueng Bien Yang Graduate : Chien-Ming Hsiao.
A Region Based Stereo Matching Algorithm Using Cooperative Optimization Zeng-Fu Wang, Zhi-Gang Zheng University of Science and Technology of China Computer.
Wenqi Zhu 3D Reconstruction From Multiple Views Based on Scale-Invariant Feature Transform.
Computer Vision Lecture #10 Hossam Abdelmunim 1 & Aly A. Farag 2 1 Computer & Systems Engineering Department, Ain Shams University, Cairo, Egypt 2 Electerical.
Scene Completion Using Millions of Photographs James Hays, Alexei A. Efros Carnegie Mellon University ACM SIGGRAPH 2007.
Robust Object Tracking by Hierarchical Association of Detection Responses Present by fakewen.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Synthesizing Natural Textures Michael Ashikhmin University of Utah.
1 Regions and Binary Images Hao Jiang Computer Science Department Sept. 24, 2009.
Course14 Dynamic Vision. Biological vision can cope with changing world Moving and changing objects Change illumination Change View-point.
Jeong Kanghun CRV (Computer & Robot Vision) Lab..
    LICENSE PLATE EXTRACTION AND CHARACTER SEGMENTATION   By HINA KOCHHAR NITI GOEL Supervisor Dr. Rajeev Srivastava        
Nottingham Image Analysis School, 23 – 25 June NITS Image Segmentation Guoping Qiu School of Computer Science, University of Nottingham
Wonjun Kim and Changick Kim, Member, IEEE
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Media Processor Lab. Media Processor Lab. Trellis-based Parallel Stereo Matching Media Processor Lab. Sejong univ.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
Digital Image Processing CCS331 Relationships of Pixel 1.
Portable Camera-Based Assistive Text and Product Label Reading From Hand-Held Objects for Blind Persons.
1 2D TO 3D IMAGE AND VIDEO CONVERSION. INTRODUCTION The goal is to take already existing 2D content, and artificially produce the left and right views.
License Plate Recognition of A Vehicle using MATLAB
April 21, 2016Introduction to Artificial Intelligence Lecture 22: Computer Vision II 1 Canny Edge Detector The Canny edge detector is a good approximation.
A Plane-Based Approach to Mondrian Stereo Matching
Digital Image Processing Lecture 15: Morphological Algorithms April 27, 2005 Prof. Charlene Tsai.
Semi-Global Matching with self-adjusting penalties
Mean Shift Segmentation
Computer Vision Lecture 5: Binary Image Processing
Scott Tan Boonping Lau Chun Hui Weng
Computer and Robot Vision I
Course 6 Stereo.
Presentation transcript:

Introduction Autostereoscopic displays give great promise as the future of 3D technology. These displays spatially multiplex many views onto a screen, providing a more immersive experience by enabling users to look around the virtual scene from different angles.Multiview screens commonly display 8 images and future screens will certainly display even more. The most difficult task for multiview imagery is how to capture content from so many cameras. Using many cameras is complex, expensive, requires bulky equipment, and is impossible in certain applications such as medicine. Capturing imagery from two cameras, however, is much simpler and can be done cheaply with a handheld device. Thus, if we can generate many views from a stereo image pair then we can circumvent the hardware problem. The goal of view synthesis is depicted in Fig. 1. Proposed Method Our method consists of four main steps: generating an initial view, refining it, filling the holes in the disparity map, then filling holes in the color image. The inputs to the algorithm are a stereo image pair and associated disparity maps, such as the dataset shown in Fig Generating the Initial View Consider the left and right cameras to be along a normalized baseline at positions 0 and 1, respectively. Let the virtual camera be at location α, 0 < α < 1. We can map the pixels from the two views into the virtual view by horizontally shifting the pixel locations by the disparity of the pixel scaled by α or 1 − α for the left and right views, respectively. We form two candidate images, T L and T R, corresponding to the left and right images I L and I R, respectively. Letting d L and d R denote the left and right disparity maps, respectively, each of size X × Y, we have where m indexes the Y rows and n indexes the X columns of the image. When a pixel at a particular location exists in one candidate view but not the other, we have the case where a pixel is occluded in one view but not the other. These holes must be filled as described in the subsequent steps of the algorithm. The outputs of this stage are shown in Fig. 3(a) and 3(b). Let d n (C ij;r, x ij ) be the set of n smallest distances from each element in C ij;r to x ij. The color class assignment c is In the case when no boundary pixels are valid, we let c be the most common color class among valid neighborhood pixels. We use k = 5 clusters in the k-means segmentation. The final estimated image is shown in Fig. 5(a). Results The algorithm generally performs well, with an average peak signal-to-noise ratio (PSNR) of dB and structural similarity index (SSIM) of The PSNR value is only 1 dB lower than the result in view synthesis algorithm, but has the same SSIM value. Further, the run time for view synthesis algorithm is 70 seconds per image, whereas the proposed method runs about 6 times faster at 12 seconds per synthesized view. Conclusions View synthesis is a viable solution for generating content for autostereoscopic multiview displays. Furthermore, synthesizing views will be absolutely essential for certain applications and for future displays that will incorporate more views. Synthesizing views from a stereo pair also has dramatic benefits when considering data transmission of multiview imagery. Dong-Yuh Song, Hsiu-Wen Ou-Yang Department of Computer Science, Chung Cheng University, Min-Hsiung Township, Chia-yi County Literature cited EFFICIENT STEREO-TO-MULTIVIEW SYNTHESIS Ankit K. Jain Lam C. Tran Ramsin Khoshabeh Truong Q. Nguyen University of California, San Diego Department of Electrical and Computer Engineering 9500 Gilman Drive, La Jolla, CA USA Fig. 2. “Art” dataset from Middlebury Stereo Database. (a)–(b) Stereo pair. (c)–(d) Disparity maps. 2. Refinement Disparity estimates along depth discontinuities are missing from the original disparity maps in Fig. 2, or are unreliable. We tag these unreliable pixels by performing edge detection on a binary image showing the mapping of the original left and right views into the synthesized viewpoint. where ⊕ denotes image dilation, I 1 (I 2 ) is the binary map containing a 1 if the corresponding pixel was copied from the right (left) image and a 0 elsewhere, and the structural element s is given by The two operands of the union are the edge maps of the left and right mappings, and the result is dilated to increase connectivity of edges and ensure that unreliable estimates are not left unfiltered. We filter the image and disparity map outputs of Section Generating the Initial View with a 5×5 median filter at every location where M is 1 shown in fig. 4. For further information Please contact A link to an online, PDF-version of the poster is nice, too. Fig 4. Outputs from steps in Section Figure 5.. (a) Final image output of steps in Section Filling the Image. (b) Final disparity map output of steps in Section Filling the Disparity Map. EFFICIENT STEREO-TO-MULTIVIEW SYNTHESIS Fig. 1. View synthesis problem. From cameras 0 and 1, how can we estimate what a camera at position α would see? T L (m, n − αd L (m, n)) = I L (m, n) T R (m,X − n + (1 − α)d R (m,X − n)) = I R (m, X − n) M = ([(I 1 ⊕ s) \ I 1 ] ∪ [(I 2 ⊕ s) \ I 2 ]) ⊕ s 4. Filling the Image Let R i be the ith contiguous region of missing pixels and let the hole under consideration be x ij, the jth missing pixel in R i. Centered around x ij, we select an N × N neighborhood N ij. We form a histogram of B bins and find the set of pixels that are in the same bin as the estimated disparity value for x ij. The set of valid pixels V ij in N ij contains those pixels that are in the same disparity level as x ij and are not holes. We form a super neighborhood S i, which is the smallest rectangle that contains all N ij. Perform k-means segmentation with k clusters on the gray scale neighborhood. Denote the set of pixels in S i assigned to color class r as K ir, 1 ≤ r ≤ k. To determine the color class to which the hole x ij belongs, we consider the valid border pixels in each color class: We will assign a color class to x ij based on the smallest median distance of K ir to x ij. We choose the median over the same number of distances n: Fig. 3. Outputs from steps in Section 3.1.