Learning to Perceive Transparency from the Statistics of Natural Scenes Anat Levin School of Computer Science and Engineering The Hebrew University of.

Slides:



Advertisements
Similar presentations
FMRI Methods Lecture 10 – Using natural stimuli. Reductionism Reducing complex things into simpler components Explaining the whole as a sum of its parts.
Advertisements

Removing blur due to camera shake from images. William T. Freeman Joint work with Rob Fergus, Anat Levin, Yair Weiss, Fredo Durand, Aaron Hertzman, Sam.
Land’s Retinex algorithm
Bayesian Belief Propagation
Image Registration  Mapping of Evolution. Registration Goals Assume the correspondences are known Find such f() and g() such that the images are best.
Scale & Affine Invariant Interest Point Detectors Mikolajczyk & Schmid presented by Dustin Lennon.
Optimizing and Learning for Super-resolution
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute.
Section 3: Appendix BP as an Optimization Algorithm 1.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Dynamic Bayesian Networks (DBNs)
1School of CS&Eng The Hebrew University
CMPUT 466/551 Principal Source: CMU
2008 SIAM Conference on Imaging Science July 7, 2008 Jason A. Palmer
Rob Fergus Courant Institute of Mathematical Sciences New York University A Variational Approach to Blind Image Deconvolution.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Belief Propagation on Markov Random Fields Aggeliki Tsoli.
Bayesian Robust Principal Component Analysis Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University January 21, 2011 Reading Group (Xinghao.
Graphical models, belief propagation, and Markov random fields 1.
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
Recovering Intrinsic Images from a Single Image 28/12/05 Dagan Aviv Shadows Removal Seminar.
Computer Vision - A Modern Approach
Problem Sets Problem Set 3 –Distributed Tuesday, 3/18. –Due Thursday, 4/3 Problem Set 4 –Distributed Tuesday, 4/1 –Due Tuesday, 4/15. Probably a total.
A Closed Form Solution to Natural Image Matting
PSY 5018H: Math Models Hum Behavior, Prof. Paul Schrater, Spring 2005 Making Decisions: Modeling perceptual decisions.
Learning Low-Level Vision William T. Freeman Egon C. Pasztor Owen T. Carmichael.
TEXTURE SYNTHESIS PEI YEAN LEE. What is texture? Images containing repeating patterns Local & stationary.
Understanding Belief Propagation and its Applications Dan Yuan June 2004.
1 Integration of Background Modeling and Object Tracking Yu-Ting Chen, Chu-Song Chen, Yi-Ping Hung IEEE ICME, 2006.
Belief Propagation Kai Ju Liu March 9, Statistical Problems Medicine Finance Internet Computer vision.
Image Deblurring with Optimizations Qi Shan Leo Jiaya Jia Aseem Agarwala University of Washington The Chinese University of Hong Kong Adobe Systems, Inc.
Deriving intrinsic images from image sequences. Yair Weiss, Presentation by Leonid Taycher.
Understanding and evaluating blind deconvolution algorithms
A Trainable Graph Combination Scheme for Belief Propagation Kai Ju Liu New York University.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Image Representation Gaussian pyramids Laplacian Pyramids
CAP5415: Computer Vision Lecture 4: Image Pyramids, Image Statistics, Denoising Fall 2006.
Mining Discriminative Components With Low-Rank and Sparsity Constraints for Face Recognition Qiang Zhang, Baoxin Li Computer Science and Engineering Arizona.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
Deriving Intrinsic Images from Image Sequences Mohit Gupta 04/21/2006 Advanced Perception Yair Weiss.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Computer Vision Michael Isard and Dimitris Metaxas.
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
Image Segmentation and Edge Detection Digital Image Processing Instructor: Dr. Cheng-Chien LiuCheng-Chien Liu Department of Earth Sciences National Cheng.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Monitoring and Enhancing Visual Features (movement, color) as a Method for Predicting Brain Activity Level -in Terms of the Perception of Pain Sensation.
Color and Brightness Constancy Jim Rehg CS 4495/7495 Computer Vision Lecture 25 & 26 Wed Oct 18, 2002.
Lecture#4 Image reconstruction
Machine Learning 5. Parametric Methods.
Lecture 04 Edge Detection Lecture 04 Edge Detection Mata kuliah: T Computer Vision Tahun: 2010.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Motion tracking TEAM D, Project 11: Laura Gui - Timisoara Calin Garboni - Timisoara Peter Horvath - Szeged Peter Kovacs - Debrecen.
Introduction of BP & TRW-S
Intrinsic images and shape refinement
Today.
Segmentation of Dynamic Scenes
Segmentation of Dynamic Scenes from Image Intensities
Machine Learning Basics
Learning to Combine Bottom-Up and Top-Down Segmentation
Segmentation of Dynamical Scenes
Analysis of Contour Motions
Expectation-Maximization & Belief Propagation
Introduction to Object Tracking
Gradient Domain Salience-preserving Color-to-gray Conversion
An Efficient Projection for L1-∞ Regularization
Presentation transcript:

Learning to Perceive Transparency from the Statistics of Natural Scenes Anat Levin School of Computer Science and Engineering The Hebrew University of Jerusalem Joint work with Assaf Zomet and Yair Weiss

Transparency

? How does our visual system choose the right decomposition ? Why not “simpler” one layer solution? Which two layers out of infinite possibilities?

Talk Outlines Motivation and previous work Our approach Results and future work

Transparency in the real world “Fashion Planet's photographers have spent the last five years working to bring you clean photographs of the windows on New York especially without the reflections that usually occur in such photography”

Transparency and shading

Transparency in human vision Metelli's conditions (Metelli 74) T-junctions, X-junctions, doubly reversing junctions (Adelson and Anandan 90, Anderson 99) Two layersOne layer Not obvious how to apply “junction catalogs” to real images.

Transparency from multiple frames Two frames with polarizer using ICA (Farid and Adelson 99, Zibulevsky 02) Multiple frames with specific motions (Irani et al. 94, Szeliski et al. 00, Weiss 01)

Shading from a single frame Retinex (Land and McCann 71). Color (Drew, Finlayson Hordley 02) Learning approach (Tappen, Freeman Adelson 02)

Talk Outlines Motivation and previous work Our approach Results and future work

Our Approach Ill-posed problem. Assume probability distribution Pr(I 1 ), Pr(I 2 ) and search for most probable solution. (ICA with a single microphone)

Statistics of natural scenes Input imagedx histogramdx Log histogram

Statistics of derivative filters Log histogram Generalized Gaussian distribution (Mallat 89, Simoncelli 95) Gaussian –x 2 –x 1/2 0 Log Probability Laplacian –|x|

Is sparsity enough? =+=+ Or:

Exactly the same derivatives exist in the single layer solution as in the two layers solution. Is sparsity enough? =+ =+ Or:

Beyond sparseness Higher order statistics of filter outputs (e.g. Portilla and Simoncelli 2000). Marginals of more complicated feature detectors (e.g. Zhu and Mumford 97, Della Pietra Della Pietra and Lafferty 96).

Corners and transparency In typical images, edges are sparse. Adding typical images is expected to increase the number of corners. Not true for white noise =+

Harris-like operator

Derivative FilterCorner Operator Corner histograms

Fitting: Derivative FilterCorner Operator Typical exponents for natural images:

Simple prior for transparency prediction The probability of a decomposition

Does this predict transparency?

How important are the statistics? Is it important that the statistics are non Gaussian? Would any cost that penalized high gradients and corners work?

The importance of being non Gaussian

The “scalar transparency” problem Consider a prior over positive scalars For which priors is the MAP solution sparse?

The “scalar transparency” problem Observation: The MAP solution is obtained with a=0, b=1 or a=1, b=0 if and only if f(x)=log P(x) is convex MAP solution: a=0, b=1MAP solution: a=0.5, b=0.5

The importance of being non Gaussian

? Can we perform a global optimization ?

Conversion to discrete MRF Local Potential- derivative filters: Pairwise Potential- pairwise approximation to the corner operator: -Enforcing integrability For the decomposition: gradient at location i

Conversion to discrete MRF Local Potential- derivative filters: Pairwise Potential- pairwise approximation to the corner operator: -Integrability enforcing For the decomposition:

possible assignments. Solution: use max-product belief propagation. The MRF has many cycles but BP works in similar problems (Freeman and Pasztor 99, Frey et al Sun et al 2002). Converges to strong local minimum (Weiss and Freeman 2001) Optimizing discrete MRF

Drawbacks of BP for this problem Large memory and time complexity. Convergence depends on update order. Discretization artifacts

Talk Outlines Motivation and previous work Our approach Results and future work

Results inputOutput layer 1 Output layer 2

Results inputOutput layer 1 Output layer 2

Future Work OriginalNon linear filter Dealing with a more complex texture +=

Future Work Dealing with a more complex texture: Use application specific priors (e.g. Manhattan World) Extend to shading and illumination. Applying other optimization methods. Learn discriminative features automatically A coarse qualitative separation.

Conclusions Natural scene statistics predict perception of transparency. First algorithm that can decompose a single image into the sum of two images.