Scene Completion Using Millions of Photographs James Hays, Alexei A. Efros Carnegie Mellon University ACM SIGGRAPH 2007.

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

Filling Algorithms Pixelwise MRFsChaos Mosaics Patch segments are pasted, overlapping, across the image. Then either: Ambiguities are removed by smoothing.
Image Quilting and Apples
James Hays and Alexei A. Efros Carnegie Mellon University CVPR IM2GPS: estimating geographic information from a single image Wen-Tsai Huang.
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
A Graph based Geometric Approach to Contour Extraction from Noisy Binary Images Amal Dev Parakkat, Jiju Peethambaran, Philumon Joseph and Ramanathan Muthuganapathy.
Data-driven methods: Texture (Sz 10.5) Cs129 Computational Photography James Hays, Brown, Spring 2011 Many slides from Alexei Efros.
SOFT SCISSORS: AN INTERACTIVE TOOL FOR REALTIME HIGH QUALITY MATTING International Conference on Computer Graphics and Interactive Techniques ACM SIGGRAPH.
Presented By: Vennela Sunnam
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
Data-driven Visual Similarity for Cross-domain Image Matching
Cambridge, Massachusetts Pose Estimation in Heavy Clutter using a Multi-Flash Camera Ming-Yu Liu, Oncel Tuzel, Ashok Veeraraghavan, Rama Chellappa, Amit.
S I E M E N S C O R P O R A T E R E S E A R C H 1 1 A Seeded Image Segmentation Framework Unifying Graph Cuts and Random Walker Which Yields A New Algorithm.
A Gimp Plugin that uses “GrabCut” to perform image segmentation
Video Inpainting Under Constrained Camera Motion Kedar A. Patwardhan, Student Member, IEEE, Guillermo Sapiro, Senior Member, IEEE, and Marcelo Bertalm.
IMAGE RESTORATION AND REALISM MILLIONS OF IMAGES SEMINAR YUVAL RADO.
ICIP 2000, Vancouver, Canada IVML, ECE, NTUA Face Detection: Is it only for Face Recognition?  A few years earlier  Face Detection Face Recognition 
Lecture 6 Image Segmentation
Recognition using Regions CVPR Outline Introduction Overview of the Approach Experimental Results Conclusion.
Small Codes and Large Image Databases for Recognition CVPR 2008 Antonio Torralba, MIT Rob Fergus, NYU Yair Weiss, Hebrew University.
CS 376b Introduction to Computer Vision 04 / 08 / 2008 Instructor: Michael Eckmann.
Image Quilting for Texture Synthesis and Transfer Alexei A. Efros1,2 William T. Freeman2.
Texture Synthesis on Surfaces Paper by Greg Turk Presentation by Jon Super.
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
WORD-PREDICTION AS A TOOL TO EVALUATE LOW-LEVEL VISION PROCESSES Prasad Gabbur, Kobus Barnard University of Arizona.
Highlights Lecture on the image part (10) Automatic Perception 16
Smart Traveller with Visual Translator. What is Smart Traveller? Mobile Device which is convenience for a traveller to carry Mobile Device which is convenience.
Opportunities of Scale Computer Vision James Hays, Brown Many slides from James Hays, Alyosha Efros, and Derek Hoiem Graphic from Antonio Torralba.
Opportunities of Scale, Part 2 Computer Vision James Hays, Brown Many slides from James Hays, Alyosha Efros, and Derek Hoiem Graphic from Antonio Torralba.
Graphcut Texture: Image and Video Synthesis Using Graph Cuts
Creating and Exploring a Large Photorealistic Virtual Space INRIA / CSAIL / Adobe First IEEE Workshop on Internet Vision, associated with CVPR 2008.
Computer vision.
Internet-scale Imagery for Graphics and Vision James Hays cs195g Computational Photography Brown University, Spring 2010.
Wavelet-Based Multiresolution Matching for Content-Based Image Retrieval Presented by Tienwei Tsai Department of Computer Science and Engineering Tatung.
Efficient Editing of Aged Object Textures By: Olivier Clément Jocelyn Benoit Eric Paquette Multimedia Lab.
ALIGNMENT OF 3D ARTICULATE SHAPES. Articulated registration Input: Two or more 3d point clouds (possibly with connectivity information) of an articulated.
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
Intelligent Scissors for Image Composition Anthony Dotterer 01/17/2006.
Poisson Image Editing & Terrain Synthesis Howard Zhou Jie Sun
80 million tiny images: a large dataset for non-parametric object and scene recognition CS 4763 Multimedia Systems Spring 2008.
EECS 274 Computer Vision Segmentation by Clustering II.
Data Extraction using Image Similarity CIS 601 Image Processing Ajay Kumar Yadav.
Eye regions localization Balázs Harangi – University of Debrecen Ciprian Pop – Technical University of Cluj-Napoca László Kovács – University of Debrecen.
Visual Data on the Internet With slides from Alexei Efros, James Hays, Antonio Torralba, and Frederic Heger : Computational Photography Jean-Francois.
Templates, Image Pyramids, and Filter Banks
2D Texture Synthesis Instructor: Yizhou Yu. Texture synthesis Goal: increase texture resolution yet keep local texture variation.
Graphcut Textures Image and Video Synthesis Using Graph Cuts
Non-Ideal Iris Segmentation Using Graph Cuts
1 Machine Vision. 2 VISION the most powerful sense.
Journal of Visual Communication and Image Representation
Image Segmentation Image segmentation (segmentace obrazu)
Scene Text Extraction Using Focus of Mobile Camera Egyul Kim, SeongHun Lee, JinHyung Kim Artificial Intelligence & Pattern Recognition Lab, KAIST, Korea.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Content-Based Image Retrieval Using Color Space Transformation and Wavelet Transform Presented by Tienwei Tsai Department of Information Management Chihlee.
Cell Segmentation in Microscopy Imagery Using a Bag of Local Bayesian Classifiers Zhaozheng Yin RI/CMU, Fall 2009.
SIGGRAPH 2007 Hui Fang and John C. Hart.  We propose an image editing system ◦ Preserve its detail and orientation by resynthesizing texture from the.
Hebrew University Image Processing Exercise Class 8 Panoramas – Stitching and Blending Min-Cut Stitching Many slides from Alexei Efros.
Technological Uncanny K. S'hell, C Kurtz, N. Vincent et E. André et M. Beugnet 1.
April 21, 2016Introduction to Artificial Intelligence Lecture 22: Computer Vision II 1 Canny Edge Detector The Canny edge detector is a good approximation.
Modeling Perspective Effects in Photographic Composition Zihan Zhou, Siqiong He, Jia Li, and James Z. Wang The Pennsylvania State University.
Graphcut Textures:Image and Video Synthesis Using Graph Cuts
Cutting Images: Graphs and Boundary Finding
Face Detection EE368 Final Project Group 14 Ping Hsin Lee
CS 4501: Introduction to Computer Vision Sparse Feature Detectors: Harris Corner, Difference of Gaussian Connelly Barnes Slides from Jason Lawrence, Fei.
Nonparametric Semantic Segmentation
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Image Segmentation Techniques
Announcements Guest lecture next Tuesday
Fourier Transform of Boundaries
Presentation transcript:

Scene Completion Using Millions of Photographs James Hays, Alexei A. Efros Carnegie Mellon University ACM SIGGRAPH 2007

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Introduction  Every once in a while, we all wish we could erase something from our photograph

Introduction  Image completion(inpainting, hole-filling)  Filling in or replacing an image region with new image data such that the modification can not be detected

Introduction  The data could have been there  The data should have been there

Introduction  The existing methods operate by extending adjacent textures and contours into the unknown region  Filling in the unknown region with content from the known parts of the input image

Introduction  The assumption is that all the necessary image data to fill in an unknown region is located somewhere else in the same image  This assumption is flawed

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Overview  We perform image completion by leveraging a massive database of images  Two compelling reasons  A region will be impossible to fill plausibly using only image data from the source image  Reusing that content would often leave obvious duplications

Overview  There are several challenges with drawing content from other images  Computational  Semantically invalid  Seamlessly

Overview  Alleviate computational and semantic  Find images depicting semantically similar scenes  Use only the best matching scenes to find patches which match the content surrounding the missing region  Seamlessly combine image regions  Graph cut segmentation  Poisson blending

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Semantic Scene Matching  Our image database  Download images in thirty Flickr.com groups  Download images based on keyword searches  Discarded duplicate images and images that are too small  Distributed among a cluster of 15 machines  Acquir about 2.3 million unique images

Semantic Scene Matching  Look for scenes which are most likely to be semantically equivalent to the image requiring completion  GIST descriptor  Augment the scene descriptor with color information of the query image down-sampled to the spatial resolution of the gist

Semantic Scene Matching  Given an input image to be hole-filled, we first compute its gist descriptor with the missing regions excluded  We calculate the SSD between the the gist of the query image and every gist in the database  The color difference is computed in the lab color space

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Local Context Matching  Having constrained our search to semantically similar scenes we can use Template matching to more precisely align

Local Context Matching  Pixel-wise alignment score  We define the local context to be all pixels within an 80 pixel radius of the hole’s boundary  This context is compared against the 200 best matching scenes  Using SSD error in lab color space

Local Context Matching  Texture similarity score  Measure coarse compatibility of the proposed fill-in region to the source image within the local context  Computed as a 5x5 median filter of image gradient magnitude at each pixel  The descriptors of the two images are compared via SSD

Local Context Matching  Composite each matching scene into the incomplete image at its best placement using a form of graph cut seam finding and standard poisson blending

Local Context Matching  Past image completion algorithms  The remaining valid pixels in an image can not changed  Our completion algorithms  Allow to remove valid pixels from the query image  But discourage the cutting of too many pixels

Local Context Matching  Past seam-finding  Minimum intensity difference between two images  Cause the seam to pass through many high frequency edges  Our seam-finding  Minimum the gradient of the image difference along the seam

Local Context Matching  We find the seam by minimizing the following cost function  : unary costs of assigning any pixel p, to a specific label L(p)  L(p) : patch or exist

Local Context Matching  For missing regions of the existing image  is a very large number  For regions of the image not covered by the scene match  is a very large number  For all other pixels   is pixel’s distance from the hole  k = 0.02

Local Context Matching  is non-zero only for immediately adjacent, 4-way connected pixels  L(p) = L(q), the cost is zero  L(p) L(q),  is the magnitude of the gradient of the SSD between the existing image and the scene match at pixels p and q

Local Context Matching  Finally we assign each composite a score  The scene matching distance  The local context matching distance  The local texture similarity distance  The cost of the graph cut  We present the user with the 20 composites with the lowest scores

Local Context Matching

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Results and Comparison

 Lucky  Find another image from the same physical location  It is not our goal to complete scenes and objects with their true selves in the database

Results and Comparison

 Failure cases : artifact

Results and Comparison  Failure cases : semantic violations

Results and Comparison  Failure cases : no object recognition

Results and Comparison  Failure cases : past methods perform well  For uniformly textured backgrounds  Our method is unlikely to find the exact same texture in another photograph

Outline  Introduction  Overview  Semantic Scene Matching  Local Context Matching  Results and Comparison  Conclusion

Conclusion  This paper  Present a new image completion algorithm powered by a huge database.  Unlike past methods that reuse visual data within the source image.  Further work  Two million images are still a tiny fraction of the high quality photograph available.  Our approach would be an attractive web-base application.

Thank you!!!