Introduction Belief propagation: known to produce accurate results for stereo processing/ motion estimation High storage requirements limit the use of.

Slides:



Advertisements
Similar presentations
Improved Census Transforms for Resource-Optimized Stereo Vision
Advertisements

Part Based Models Andrew Harp. Part Based Models Physical arrangement of features.
Spatial-Temporal Consistency in Video Disparity Estimation ICASSP 2011 Ramsin Khoshabeh, Stanley H. Chan, Truong Q. Nguyen.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu Toward Optimal Deployment of Communication-Intensive Cloud Applications 1.
Mixed-Resolution Patch- Matching (MRPM) Harshit Sureka and P.J. Narayanan (ECCV 2012) Presentation by Yaniv Romano 1.
Real-Time Accurate Stereo Matching using Modified Two-Pass Aggregation and Winner- Take-All Guided Dynamic Programming Xuefeng Chang, Zhong Zhou, Yingjie.
Does Color Really Help in Dense Stereo Matching?
SWE 423: Multimedia Systems
Yung-Lin Huang, Yi-Nung Liu, and Shao-Yi Chien Media IC and System Lab Graduate Institute of Networking and Multimedia National Taiwan University Signal.
A New Block Based Motion Estimation with True Region Motion Field Jozef Huska & Peter Kulla EUROCON 2007 The International Conference on “Computer as a.
2009/04/07 Yun-Yang Ma.  Overview  What is CUDA ◦ Architecture ◦ Programming Model ◦ Memory Model  H.264 Motion Estimation on CUDA ◦ Method ◦ Experimental.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
Optimizing and Auto-Tuning Belief Propagation on the GPU Scott Grauer-Gray and Dr. John Cavazos Computer and Information Sciences, University of Delaware.
Order-Independent Texture Synthesis Li-Yi Wei Marc Levoy Gcafe 1/30/2003.
1 Abstract This paper presents a novel modification to the classical Competitive Learning (CL) by adding a dynamic branching mechanism to neural networks.
DCABES 2009 China University Of Geosciences 1 The Parallel Models of Coronal Polarization Brightness Calculation Jiang Wenqian.
Feature Screening Concept: A greedy feature selection method. Rank features and discard those whose ranking criterions are below the threshold. Problem:
Cliff Rhyne and Jerry Fu June 5, 2007 Parallel Image Segmenter CSE 262 Spring 2007 Project Final Presentation.
Real-Time Stereo Matching on Programmable Graphics Hardware Liang Wei.
A Performance and Energy Comparison of FPGAs, GPUs, and Multicores for Sliding-Window Applications From J. Fowers, G. Brown, P. Cooke, and G. Stitt, University.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Image Analogies Aaron Hertzmann (1,2) Charles E. Jacobs (2) Nuria Oliver (2) Brian Curless (3) David H. Salesin (2,3) 1 New York University 1 New York.
Image Registration of Very Large Images via Genetic Programming Sarit Chicotay Omid E. David Nathan S. Netanyahu CVPR ‘14 Workshop on Registration of Very.
Reconstructing Relief Surfaces George Vogiatzis, Philip Torr, Steven Seitz and Roberto Cipolla BMVC 2004.
Introduction Autostereoscopic displays give great promise as the future of 3D technology. These displays spatially multiplex many views onto a screen,
: Chapter 12: Image Compression 1 Montri Karnjanadecha ac.th/~montri Image Processing.
Automatic Registration of Color Images to 3D Geometry Computer Graphics International 2009 Yunzhen Li and Kok-Lim Low School of Computing National University.
Rendering Adaptive Resolution Data Models Daniel Bolan Abstract For the past several years, a model for large datasets has been developed and extended.
A Local Adaptive Approach for Dense Stereo Matching in Architectural Scene Reconstruction C. Stentoumis 1, L. Grammatikopoulos 2, I. Kalisperakis 2, E.
Rohith MV, Gowri Somanath, Chandra Kambhamettu Video/Image Modeling and Synthesis(VIMS) Lab, Dept. of Computer and Information Sciences Cathleen Geiger.
Takuya Matsuo, Norishige Fukushima and Yutaka Ishibashi
Avoiding Segmentation in Multi-digit Numeral String Recognition by Combining Single and Two-digit Classifiers Trained without Negative Examples Dan Ciresan.
Scientific Writing Abstract Writing. Why ? Most important part of the paper Number of Readers ! Make people read your work. Sell your work. Make your.
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 34, NO. 2, FEBRUARY Leonardo De-Maeztu, Arantxa Villanueva, Member, IEEE, and.
Disclosure risk when responding to queries with deterministic guarantees Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Low-Power H.264 Video Compression Architecture for Mobile Communication Student: Tai-Jung Huang Advisor: Jar-Ferr Yang Teacher: Jenn-Jier Lien.
Feature-Based Stereo Matching Using Graph Cuts Gorkem Saygili, Laurens van der Maaten, Emile A. Hendriks ASCI Conference 2011.
A Region Based Stereo Matching Algorithm Using Cooperative Optimization Zeng-Fu Wang, Zhi-Gang Zheng University of Science and Technology of China Computer.
Stereo Matching Using Loopy Belief Propagation Li Zhang and Lin Liao April 23, 2004.
Window-based Approach For Fast Stereo Correspondence Raj Kumar Gupta, Siu-Yeung Cho IET Computer Vision,
SIGGRAPH 2011 ASIA Preview Seminar Rendering: Accuracy and Efficiency Shinichi Yamashita Triaxis Co.,Ltd.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology Fast accurate fuzzy clustering through data reduction Advisor.
Learning from Positive and Unlabeled Examples Investigator: Bing Liu, Computer Science Prime Grant Support: National Science Foundation Problem Statement.
Fast Census Transform-based Stereo Algorithm using SSE2
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Spatiotemporal Saliency Map of a Video Sequence in FPGA hardware David Boland Acknowledgements: Professor Peter Cheung Mr Yang Liu.
Su-ting, Chuang 1. Outline Introduction Related work Hardware configuration Detection system Optimal parameter estimation framework Conclusion 2.
Improved Census Transforms for Resource-Optimized Stereo Vision
Jeong Kanghun CRV (Computer & Robot Vision) Lab..
Visual Odometry David Nister, CVPR 2004
Collaborative Filtering via Euclidean Embedding M. Khoshneshin and W. Street Proc. of ACM RecSys, pp , 2010.
3-D WAVELET BASED VIDEO CODER By Nazia Assad Vyshali S.Kumar Supervisor Dr. Rajeev Srivastava.
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Canny Edge Detection Using an NVIDIA GPU and CUDA Alex Wade CAP6938 Final Project.
Hierarchical Systolic Array Design for Full-Search Block Matching Motion Estimation Noam Gur Arie,August 2005.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
1 M15338: Depth Map Estimation Software version 2 April, 27th 2008, Archamps Olgierd Stankiewicz Krzysztof Wegner team supervisor: Marek Domański Chair.
1שידור ווידיאו ואודיו ברשת האינטרנט Dr. Ofer Hadar Communication Systems Engineering Department Ben-Gurion University of the Negev URL:
Visual Tracking of Surgical Tools in Retinal Surgery using Particle Filtering Group 14 William Yang and David Li Presenter: William Yang Mentor: Dr. Rogerio.
Dr. Ofer Hadar Communication Systems Engineering Department
Discrete ABC Based on Similarity for GCP
Summary of “Efficient Deep Learning for Stereo Matching”
Semi-Global Matching with self-adjusting penalties
Paper – Stephen Se, David Lowe, Jim Little
SoC and FPGA Oriented High-quality Stereo Vision System
Finite Element Surface-Based Stereo 3D Reconstruction
Pei Fan*, Ji Wang, Zibin Zheng, Michael R. Lyu
Presentation transcript:

Introduction Belief propagation: known to produce accurate results for stereo processing/ motion estimation High storage requirements limit the use of the algorithm in some situations Belief propagation is an iterative algorithm which has been shown to produce accurate results for stereo processing and motion estimation via the use of a ‘data cost’ for each possible disparity / motion at each pixel and a ‘discontinuity cost’ between the results at each pair of neighboring pixels. The algorithm is used in 3 of the top 5 best-performing stereo algorithms according to the Middlebury benchmark. In stereo processing / motion estimation, belief propagation operates via the utilization of the data costs together with four sets of frequently-updated message values for each possible disparity / motion vector at each pixel, necessitating storage space of O(5 * image height * image width * number of possible disparity values / motion vectors). This storage requirement limits the use of the algorithm in some situations, particularly in cases with a large range of possible output disparities or motion vectors, which may include results at the sub-pixel level. Hierarchical Belief Propagation: Reduces the storage requirements of the algorithm Hierarchical belief propagation as presented by Felzenswalb and Huttenlocher [1] has been used to decrease the number of iterations needed for message value convergence and therefore speed up various implementations of belief propagation. However, this method does not reduce the search space or the storage requirements of the implementation. In our implementation, we use ‘rough’ results computed at the higher levels of the hierarchy to initialize the search space at the lower levels, eliminating the need to consider every possible result at each level in the hierarchy. Acknowledgments This work was partially supported by a U.S. National Aeronautics and Space Administration award NASA NNX08AD80G under the ROSES Applied Information Systems Research program Conclusions and Future Work We have presented a method to reduce the storage requirements of a stereo or motion estimation belief propagation implementation without sacrificing much in terms of the accuracy of the results, as well as how our GPU implementation can be used without modification to retrieve sub-pixel accuracy in the resulting disparity map or set of motion vectors. For image sets with a large disparity/motion range or with sub-pixel accuracy requirements, our implementation may be used where traditional belief propagation is not feasible due to overwhelming storage requirements. Future work includes further tweaking of the parameters and algorithm details to retrieve the optimal balance of storage requirements and running time versus the accuracy of the results for various problems on a variety of hardware, as well as investigating how this implementation could fit into related work that prominently incorporates belief propagation, including some of the top-scoring algorithms in the Middlebury stereo evaluation. Scott Grauer-Gray, Chandra Kambhamettu Literature Cited on Poster [1] P. F. Felzenszwalb and D. P. Huttenlocher. Efficient belief propagation for early vision. Int. J. Comput. Vision, 70(1):41-54, [2] S. Grauer-Gray, C. Kambhamettu, and K. Palaniappan. GPU implementation of belief propagation using cuda for cloud tracking and reconstruction. In 2008 IAPR Workshop on Pattern Recognition in Remote Sensing (PRRS 2008), pages 1-4, See paper for additional citations. Hierarchical Belief Propagation To Reduce Search Space Using CUDA for Stereo and Motion Estimation CUDA Implementation: Leverages the parallel processing and interpolation capabilities of the graphics processing unit (GPU) Our belief propagation implementation builds on our previous work in [2] which used the graphics processing unit (GPU) and the CUDA technology available from nVidia; the results in that work showed that the architecture can be used to speed up an implementation of the algorithm without losing accuracy, largely due to the natural parallelism of many steps of the algorithm. In addition, our implementation takes advantage of the interpolation capabilities of the GPU to retrieve sub-pixel disparities without making changes to the general implementation. We ran our implementation on the Tsukuba stereo set where all the ground-truth disparities are integer values, using a disparity range from 0 to 16. In our initial trial, we ran the implementation on a single level with a disparity increment of 1, resulting in a search space of 17 possible disparities values. In the next trial, we ran our implementation using 2 levels, where the increment in disparity candidates is set to 2 at the upper level and halved to 1 in the lower level, resulting in a search space of 9 possibly disparities at each level and nearly halving the storage requirements of the implementation as compared to the single-level trial. We continued to run our implementation on the Tsukuba set using 3 and 4 levels, with disparity candidate increments starting at 4 and 8 that are halved in each succeeding level, resulting in search spaces of 5 and 3 possible disparities and significantly lower storage requirements. The Tsukuba reference image, ground truth disparity map, and the resulting disparity maps from our trials are shown in figure 1 below, and we provide quantitative statistics regarding the accuracy of our implementation for each trial in table 1 below the figure. These results show that our implementation is capable of reducing the storage requirements of the belief propagation algorithm without severely sacrificing the accuracy of the results in this image set. Next, we ran our implementation on the ‘Venus” stereo set with ground truth disparities ranging from 3.0 to that are given at increments of one-eighth of a pixel. Our implementation takes advantage of the interpolation capabilities of the GPU, treating a non-integer disparity candidate the same way as an integer disparity candidate without any change in implementation via retrieval from texture memory on the hardware. We ran our implementation on the stereo set at a single level using disparity increments ranging from 0.4 to 1.0 with a disparity candidate range from 0.0 to The reference ‘Venus’ image, ground truth disparity map, and the resulting disparity maps using disparity increments of 0.5 and 1.0 are shown in figure 2 below, and the quantitative results showing the percentage of ‘bad’ pixels beyond an error threshold of 0.5 and 1.0 are shown in table 2 to the right of the figure. Results On a Stereo Set with Sub-Pixel Disparities Belief propagation has been shown to give accurate results for motion estimation as well as stereo, but the presence of a two dimensional motion search space which is often combined with a need for sub-pixel accuracy can lead to large and possibly infeasible storage requirements when attempting traditional belief propagation for motion estimation. Our hierarchical implementation can make motion estimation via belief propagation feasible by reducing the storage requirements. We ran our implementation on the Dimetrodon image pair using a motion range from -5.0 to 5.0 on the x and y axis and with a motion increment of 2.5 at the top level. The motion increment is halved with each succeeding level of the hierarchy. The first frame of the image pair, the legend showing color coding of motion, the ground truth motion, and the computed motion using our implementation with 5 levels are shown in figure 3 below. The table to the right of the figure displays quantitative results showing the accuracy of our implementation using 3, 4, and 5 levels as compared to the ground truth motion. Results On a Set of Sequential Images Containing 2D Motion Storage requirements reduced with minimal sacrifice of accuracy of resultant disparity maps Results On a Stereo Set with Integer Disparities Accuracy of resultant disparity map increases when non-integer disparity values are included in search space Motion estimation using belief propagation more feasible via lower storage requirements University of Delaware, Newark, DE Video/Image Modeling and Synthesis Lab (VIMS), Department of Computer and Information Sciences,