Analysis of Contour Motions

Slides:



Advertisements
Similar presentations
Human-Assisted Motion Annotation Ce Liu William T. Freeman Edward H. Adelson Massachusetts Institute of Technology Yair Weiss The Hebrew University of.
Advertisements

Bayesian Belief Propagation
Probabilistic Tracking and Recognition of Non-rigid Hand Motion
CISC 489/689 Spring 2009 University of Delaware
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute.
Dynamic Occlusion Analysis in Optical Flow Fields
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
Computer Vision Laboratory 1 Unrestricted Recognition of 3-D Objects Using Multi-Level Triplet Invariants Gösta Granlund and Anders Moe Computer Vision.
Uncertainty Representation. Gaussian Distribution variance Standard deviation.
Computer Vision Optical Flow
Formation et Analyse d’Images Session 8
A New Block Based Motion Estimation with True Region Motion Field Jozef Huska & Peter Kulla EUROCON 2007 The International Conference on “Computer as a.
Hierarchical Region-Based Segmentation by Ratio-Contour Jun Wang April 28, 2004 Course Project of CSCE 790.
Announcements Quiz Thursday Quiz Review Tomorrow: AV Williams 4424, 4pm. Practice Quiz handout.
Problem Sets Problem Set 3 –Distributed Tuesday, 3/18. –Due Thursday, 4/3 Problem Set 4 –Distributed Tuesday, 4/1 –Due Tuesday, 4/15. Probably a total.
Segmentation Divide the image into segments. Each segment:
Detecting and Tracking Moving Objects for Video Surveillance Isaac Cohen and Gerard Medioni University of Southern California.
High-Quality Video View Interpolation
Segmentation by Clustering Reading: Chapter 14 (skip 14.5) Data reduction - obtain a compact representation for interesting image data in terms of a set.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Today Introduction to MCMC Particle filters and MCMC
1 Stanford CS223B Computer Vision, Winter 2006 Lecture 7 Optical Flow Professor Sebastian Thrun CAs: Dan Maynes-Aminzade, Mitul Saha, Greg Corrado Slides.
Matching Compare region of image to region of image. –We talked about this for stereo. –Important for motion. Epipolar constraint unknown. But motion small.
Automatic Estimation and Removal of Noise from a Single Image
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Optical Flow Donald Tanguay June 12, Outline Description of optical flow General techniques Specific methods –Horn and Schunck (regularization)
Intelligent Vision Systems ENT 496 Object Shape Identification and Representation Hema C.R. Lecture 7.
#MOTION ESTIMATION AND OCCLUSION DETECTION #BLURRED VIDEO WITH LAYERS
Motion Segmentation By Hadas Shahar (and John Y.A.Wang, and Edward H. Adelson, and Wikipedia and YouTube) 1.
Visual motion Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Markov Random Fields Probabilistic Models for Images
Learning the Appearance and Motion of People in Video Hedvig Sidenbladh, KTH Michael Black, Brown University.
2/14/00 Computer Vision. 2/14/00 Computer Vision Lecturer: Ir. Resmana Lim, M.Eng. Text: 1) Computer Vision -- A Modern Approach.
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
A DISTRIBUTION BASED VIDEO REPRESENTATION FOR HUMAN ACTION RECOGNITION Yan Song, Sheng Tang, Yan-Tao Zheng, Tat-Seng Chua, Yongdong Zhang, Shouxun Lin.
Optical Flow. Distribution of apparent velocities of movement of brightness pattern in an image.
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
Features, Feature descriptors, Matching Jana Kosecka George Mason University.
Speaker Min-Koo Kang March 26, 2013 Depth Enhancement Technique by Sensor Fusion: MRF-based approach.
Representing Moving Images with Layers J. Y. Wang and E. H. Adelson MIT Media Lab.
Optical flow and keypoint tracking Many slides adapted from S. Seitz, R. Szeliski, M. Pollefeys.
Motion Segmentation at Any Speed Shrinivas J. Pundlik Department of Electrical and Computer Engineering, Clemson University, Clemson, SC.
MASKS © 2004 Invitation to 3D vision Lecture 3 Image Primitives andCorrespondence.
11/25/03 3D Model Acquisition by Tracking 2D Wireframes Presenter: Jing Han Shiau M. Brown, T. Drummond and R. Cipolla Department of Engineering University.
Motion and Optical Flow
Robust Visual Motion Analysis: Piecewise-Smooth Optical Flow
LOCUS: Learning Object Classes with Unsupervised Segmentation
Nonparametric Semantic Segmentation
Image Primitives and Correspondence
Dynamical Statistical Shape Priors for Level Set Based Tracking
Outline Image Segmentation by Data-Driven Markov Chain Monte Carlo
Detecting Artifacts and Textures in Wavelet Coded Images
Segmentation and Grouping
Representing Moving Images with Layers
Representing Moving Images with Layers
Motion Estimation Today’s Readings
Parallel Integration of Video Modules
PRAKASH CHOCKALINGAM, NALIN PRADEEP, AND STAN BIRCHFIELD
Announcements more panorama slots available now
EM for Motion Segmentation
Announcements Questions on the project? New turn-in info online
Coupled Horn-Schunck and Lukas-Kanade for image processing
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Introduction to Object Tracking
Announcements more panorama slots available now
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
Optical flow and keypoint tracking
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Analysis of Contour Motions Neural Information Processing Systems 2006 Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology

Visual Motion Analysis in Computer Vision Motion analysis is essential in Video processing Geometry reconstruction Object tracking, segmentation and recognition Graphics applications Is motion analysis solved? Do we have good representation for motion analysis? Is it computationally feasible to infer the representation from the raw video data? What is a good representation for motion? Motion analysis is the corner stone for many computer vision applications.

Seemingly Simple Examples Let’s look at these two seemingly simple examples. We can clearly see a moving square occluding four circles from the Kanizsa square example, and one leg moving occluding the other from the real video. These two examples both contain textueless objects with occlusions. What do you expect the output from the state-of-the-art motion analysis algorithms for these two examples? Kanizsa square From real video

Output from the State-of-the-Art Optical Flow Algorithm Kanizsa square Optical flow field T. Brox et al. High accuracy optical flow estimation based on a theory for warping. ECCV 2004

Output from the State-of-the-Art Optical Flow Algorithm Dancer Optical flow field T. Brox et al. High accuracy optical flow estimation based on a theory for warping. ECCV 2004

Optical flow representation: aperture problem How computer sees the motion in flow estimation? Corners Lines Flat regions Spurious junctions Boundary ownership Illusory boundaries

Optical Flow Representation We need motion representation beyond pixel level! Corners Lines Flat regions Spurious junctions Boundary ownership Illusory boundaries

Layer Representation Video is a composite of layers Layer segmentation assumes sufficient textures for each layer to represent motion A true success? J. Wang & E. H. Adelson 1994 Y. Weiss & E. H. Adelson 1994 Achieved with the help of spatial segmentation

Layer Representation Video is a composite of layers Layer segmentation assumes sufficient textures for each layer to represent motion A true success? J. Wang & E. H. Adelson 1994 Y. Weiss & E. H. Adelson 1994 Layer representation is good, but the existing layer segmentation algorithms cannot find the right layers for textureless objects Achieved with the help of spatial segmentation

Challenge: Textureless Objects under Occlusion Corners are not always trustworthy (junctions) Flat regions do not always move smoothly (discontinuous at illusory boundaries) How about boundaries? Easy to detect and track for textureless objects Able to handle junctions with illusory boundaries

Analysis of Contour Motions Our approach: simultaneous grouping and motion analysis Multi-level contour representation Junctions are appropriated handled Formulate graphical model that favors good contour and motion criteria Inference using importance sampling Contribution An important component in motion analysis toolbox for textureless objects under occlusion

Three Levels of Contour Representation Edgelets: edge particles Boundary fragments: a chain of edgelets with small curvatures Contours: a chain of boundary fragments Forming boundary fragments: easy (for textureless objects) Forming contours: hard (the focus of our work)

Overview of our system 1. Extract boundary fragments 2. Edgelet tracking with uncertainty. 3. Boundary grouping and illusory boundary 4. Motion estimation based on the grouping

Forming Boundary Fragments Boundary fragments extraction in frame 1 Steerable filters to obtain edge energy for each orientation band Spatially trace boundary fragments Boundary fragments: lines or curves with small curvature Temporal edgelet tracking with uncertainties Frame 1: edgelet (x, y, q) Frame 2: orientation energy of q A Gaussian pdf is fit with the weight of orientation energy 1D uncertainty of motion (even for T-junctions) (a) (b) (c) (d)

Forming Contours: Boundary Fragments Grouping Grouping representation: switch variables (attached to every end of the fragments) Exclusive: one end connects to at most one other end Reversible: if end (i,ti) connects to (j,tj), then (j,tj) connects to (i,ti) 1 Arbitrarily possible connection A legal contour grouping Reversibility Another legal contour grouping 1 1

Local Spatial-Temporal Cues for Grouping Illusory boundaries corresponding to the groupings (generated by spline interpolation) Motion stimulus

Local spatial-temporal cues for grouping: (a) Motion similarity The grouping with higher motion similarity is favored Velocity space KL( ) < KL( ) Motion stimulus

Local spatial-temporal cues for grouping: (b) Curve smoothness The grouping with smoother and shorter illusory boundary is favored Motion stimulus

Local spatial-temporal cues for grouping: (c) Contrast consistency The grouping with consistent local contrast is favored Motion stimulus

The Graphical Model for Grouping Affinity metric terms (a) Motion similarity (b) Curve smoothness (c) Contrast consistency The graphical model for grouping affinity reversibility no self-intersection

Motion estimation for grouped contours Gaussian MRF (GMRF) within a boundary fragment The motions of two end edgelets are similar if they are grouped together The graphical model of motion: joint Gaussian given the grouping This problem is solved in early work: Y. Weiss, Interpreting images by propagating Bayesian beliefs, NIPS, 1997.

Inference Two-step inference Grouping (switch variables) Motion based on grouping (easy, least square) Grouping: importance sampling to estimate the marginal of the switch variables Bidirectional proposal density Toss the sample if self-intersection is detected Obtain the optimal grouping from the marginal

Why bidirectional proposal in sampling?

Why bidirectional proposal in sampling? Affinity metric of the switch variable (darker, thicker means larger affinity) b1b2: 0.39 b1b3: 0.01 b1b4: 0.60 b4b1: 0.20 b4b2: 0.05 b4b3: 0.85 b2b1: 0.50 b2b3: 0.45 b2b4: 0.05 b3b1: 0.01 b3b2: 0.45 b3b4: 0.54 b1b2: 0.1750 b1b3: 0.0001 b1b4: 0.1200 Normalized affinity metrics Bidirectional proposal

Why bidirectional proposal in sampling? Bidirectional proposal of the switch variable (darker, thicker means larger affinity) b1b2: 0.39 b1b3: 0.01 b1b4: 0.60 b4b1: 0.20 b4b2: 0.05 b4b3: 0.85 b2b1: 0.50 b2b3: 0.45 b2b4: 0.05 b3b1: 0.01 b3b2: 0.45 b3b4: 0.54 b1b2: 0.62 b1b3: 0.00 b1b4: 0.38 Normalized affinity metrics Bidirectional proposal (Normalized)

Example of Sampling Motion stimulus Self intersection

Example of Sampling Motion stimulus A valid grouping

Example of Sampling Motion stimulus More valid groupings

Example of Sampling Motion stimulus More valid groupings

From Affinity to Marginals Affinity metric of the switch variable (darker, thicker means larger affinity) Motion stimulus

From Affinity to Marginals Marginal distribution of the switch variable (darker, thicker means larger affinity) Greedy algorithm to search for the best grouping based on the marginals Motion stimulus

Experiments All the results are generated using the same parameter settings Running time depends on the number of boundary fragments, varying from ten seconds to a few minutes in MATLAB

Two Moving Bars Frame 1

Two Moving Bars Frame 2

Two Moving Bars Extracted boundary fragments. The green circles are the boundary fragment end points.

Two Moving Bars Optical flow from Lucas-Kanade algorithm. The flow vectors are only plotted at the edgelets

Estimated motion by our system after grouping Two Moving Bars Estimated motion by our system after grouping

Two Moving Bars Boundary grouping and illusory boundaries (frame 1). The fragments belonging to the same contour are plotted in one color.

Two Moving Bars Boundary grouping and illusory boundaries (frame 2). The fragments belonging to the same contour are plotted in one color.

Kanizsa Square

Frame 1

Frame 2

Extracted boundary fragments

Optical flow from Lucas-Kanade algorithm

Estimated motion by our system, after grouping

Boundary grouping and illusory boundaries (frame 1)

Boundary grouping and illusory boundaries (frame 2)

Dancer

Frame 1

Frame 2

Extracted boundary fragments

Optical flow from Lucas-Kanade algorithm

Estimated motion by our system, after grouping

Lucas-Kanade flow field Estimated motion by our system, after grouping

Boundary grouping and illusory boundaries (frame 1)

Boundary grouping and illusory boundaries (frame 2)

Rotating Chair

Frame 1

Frame 2

Extracted boundary fragments

Estimated flow field from Brox et al.

Estimated motion by our system, after grouping

Boundary grouping and illusory boundaries (frame 1)

Boundary grouping and illusory boundaries (frame 2)

Conclusion A contour-based representation to estimate motion for textureless objects under occlusion Motion ambiguities are preserved and resolved through appropriate contour grouping An important component in motion analysis toolbox To be combined with the classical motion estimation techniques to analyze complex scenes

Thanks! Analysis of Contour Motions Ce Liu William T. Freeman Edward H. Adelson Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology http://people.csail.mit.edu/celiu/contourmotions/

Backup Slides

Why bidirectional proposal in sampling?

Why bidirectional proposal in sampling? Affinity metric of the switch variable (darker, thicker means larger affinity) b1b2: 0.39 b1b3: 0.01 b1b4: 0.60 b4b1: 0.20 b4b2: 0.05 b4b3: 0.85 b2b1: 0.50 b2b3: 0.45 b2b4: 0.05 b3b1: 0.01 b3b2: 0.45 b3b4: 0.54 b1b2: 0.1750 b1b3: 0.0001 b1b4: 0.1200 Normalized affinity metrics Bidirectional proposal

Why bidirectional proposal in sampling? Bidirectional proposal of the switch variable (darker, thicker means larger affinity) b1b2: 0.39 b1b3: 0.01 b1b4: 0.60 b4b1: 0.20 b4b2: 0.05 b4b3: 0.85 b2b1: 0.50 b2b3: 0.45 b2b4: 0.05 b3b1: 0.01 b3b2: 0.45 b3b4: 0.54 b1b2: 0.62 b1b3: 0.00 b1b4: 0.38 Normalized affinity metrics Bidirectional proposal (Normalized)

Sampling Grouping (Switch Variables) Motion stimulus

Lucas-Kanade flow field Estimated motion by our system, after grouping