Video Texture 15-463: Computational Photography Alexei Efros, CMU, Fall 2006 © A.A. Efros.

Slides:



Advertisements
Similar presentations
Filling Algorithms Pixelwise MRFsChaos Mosaics Patch segments are pasted, overlapping, across the image. Then either: Ambiguities are removed by smoothing.
Advertisements

Data-driven methods: Texture (Sz 10.5) Cs129 Computational Photography James Hays, Brown, Spring 2011 Many slides from Alexei Efros.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
Actions in video Monday, April 25 Kristen Grauman UT-Austin.
Introduction To Tracking
Generation of Virtual Image from Multiple View Point Image Database Haruki Kawanaka, Nobuaki Sado and Yuji Iwahori Nagoya Institute of Technology, Japan.
Real-Time Human Pose Recognition in Parts from Single Depth Images Presented by: Mohammad A. Gowayyed.
Computational Photography and Video: Video Synthesis Prof. Marc Pollefeys.
More details on presentations Aim to speak for ~50 min (after 15 min review, leaving 10 min for discussions) Try to plan discussion topics It’s fine to.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Introduction to Data-driven Animation Jinxiang Chai Computer Science and Engineering Texas A&M University.
Boundary matting for view synthesis Samuel W. Hasinoff Sing Bing Kang Richard Szeliski Computer Vision and Image Understanding 103 (2006) 22–32.
Advanced Computer Graphics (Fall 2010) CS 283, Lecture 24: Motion Capture Ravi Ramamoorthi Most slides courtesy.
Image Quilting for Texture Synthesis & Transfer Alexei Efros (UC Berkeley) Bill Freeman (MERL) +=
A Study of Approaches for Object Recognition
Direct Methods for Visual Scene Reconstruction Paper by Richard Szeliski & Sing Bing Kang Presented by Kristin Branson November 7, 2002.
Detecting and Tracking Moving Objects for Video Surveillance Isaac Cohen and Gerard Medioni University of Southern California.
Image Manifolds : Learning-based Methods in Vision Alexei Efros, CMU, Spring 2007 © A.A. Efros With slides by Dave Thompson.
Background Subtraction and Matting : Computational Photography Alexei Efros, CMU, Fall 2006 © Yuri Bonder.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2005 with a lot of slides stolen from Steve Seitz and.
High-Quality Video View Interpolation
Recognizing and Tracking Human Action Josephine Sullivan and Stefan Carlsson.
Automatic Image Alignment (feature-based) : Computational Photography Alexei Efros, CMU, Fall 2006 with a lot of slides stolen from Steve Seitz and.
Data-driven methods: Video : Computational Photography Alexei Efros, CMU, Fall 2007 © A.A. Efros.
Face Recognition and Retrieval in Video Basic concept of Face Recog. & retrieval And their basic methods. C.S.E. Kwon Min Hyuk.
Texture Optimization for Example-based Synthesis
The Uncanny Valley in 3D Modeling Steve Seitz and Rick Szeliski BIRS Workshop on Computer Vision and the Internet August 31, 2009.
ICCV 2003UC Berkeley Computer Vision Group Recognizing Action at a Distance A.A. Efros, A.C. Berg, G. Mori, J. Malik UC Berkeley.
Recognizing Action at a Distance A.A. Efros, A.C. Berg, G. Mori, J. Malik UC Berkeley.
Texture Synthesis by Non-parametric Sampling Alexei Efros and Thomas Leung UC Berkeley.
Computer vision.
Flow Based Action Recognition Papers to discuss: The Representation and Recognition of Action Using Temporal Templates (Bobbick & Davis 2001) Recognizing.
Action recognition with improved trajectories
TP15 - Tracking Computer Vision, FCUP, 2013 Miguel Coimbra Slides by Prof. Kristen Grauman.
Video Textures Arno Schödl Richard Szeliski David Salesin Irfan Essa Microsoft Research, Georgia Tech.
Learning and Recognizing Human Dynamics in Video Sequences Christoph Bregler Alvina Goh Reading group: 07/06/06.
Recognizing Human Figures and Actions Greg Mori Simon Fraser University.
Activity Detection Seminar Sivan Edri.  This capability of the human vision system argues for recognition of movement directly from the motion itself,
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Data-driven methods: Video & Texture Cs195g Computational Photography James Hays, Brown, Spring 2010 Many slides from Alexei Efros.
#MOTION ESTIMATION AND OCCLUSION DETECTION #BLURRED VIDEO WITH LAYERS
Previous lecture Texture Synthesis Texture Transfer + =
Recognizing Action at a Distance Alexei A. Efros, Alexander C. Berg, Greg Mori, Jitendra Malik Computer Science Division, UC Berkeley Presented by Pundik.
Multimedia Programming 21: Video and its applications Departments of Digital Contents Sang Il Park Lots of Slides are stolen from Alexei Efros, CMU.
Advanced Computer Graphics (Fall 2010) CS 283, Lecture 26 Texture Synthesis Ravi Ramamoorthi Slides, lecture.
Action and Gait Recognition From Recovered 3-D Human Joints IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS— PART B: CYBERNETICS, VOL. 40, NO. 4, AUGUST.
1 Motion Analysis using Optical flow CIS601 Longin Jan Latecki Fall 2003 CIS Dept of Temple University.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Video Textures This is a very cool paper.. Why video texturing? Static images are boring Video clips are finite in length We need something more interesting.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
 Present by 陳群元.  Introduction  Previous work  Predicting motion patterns  Spatio-temporal transition distribution  Discerning pedestrians  Experimental.
Active Frame Selection for Label Propagation in Videos Sudheendra Vijayanarasimhan and Kristen Grauman Department of Computer Science, University of Texas.
Motion Estimation using Markov Random Fields Hrvoje Bogunović Image Processing Group Faculty of Electrical Engineering and Computing University of Zagreb.
Motion Estimation Today’s Readings Trucco & Verri, 8.3 – 8.4 (skip 8.3.3, read only top half of p. 199) Newton's method Wikpedia page
Image-Based Rendering Geometry and light interaction may be difficult and expensive to model –Think of how hard radiosity is –Imagine the complexity of.
Data Driven Models of Motion Walking the Fine Line Between Performance, Realism and Style Chris White G-Lunch 2007.
Video Textures Arno Schödl Richard Szeliski David Salesin Irfan Essa Microsoft Research, Georgia Tech.
Visual Recognition Tutorial1 Markov models Hidden Markov models Forward/Backward algorithm Viterbi algorithm Baum-Welch estimation algorithm Hidden.
Graphcut Textures:Image and Video Synthesis Using Graph Cuts
Steps Towards the Convergence of Graphics, Vision, and Video
Motion and Optical Flow
Dynamical Statistical Shape Priors for Level Set Based Tracking
Texture Synthesis by Non-parametric Sampling
Video-based human motion recognition using 3D mocap data
Data-driven methods: Video Textures
Image Quilting for Texture Synthesis & Transfer
WELCOME.
CSSE463: Image Recognition Day 30
CSSE463: Image Recognition Day 30
Presentation transcript:

Video Texture : Computational Photography Alexei Efros, CMU, Fall 2006 © A.A. Efros

Weather Forecasting for Dummies™ Let’s predict weather: Given today’s weather only, we want to know tomorrow’s Suppose weather can only be {Sunny, Cloudy, Raining} The “Weather Channel” algorithm: Over a long period of time, record: –How often S followed by R –How often S followed by S –Etc. Compute percentages for each state: –P(R|S), P(S|S), etc. Predict the state with highest probability! It’s a Markov Chain

Markov Chain What if we know today and yestarday’s weather?

Text Synthesis [ Shannon,’48] proposed a way to generate English-looking text using N-grams: Assume a generalized Markov model Use a large text to compute prob. distributions of each letter given N-1 previous letters Starting from a seed repeatedly sample this Markov chain to generate new letters Also works for whole words WE NEEDTOEATCAKE

Mark V. Shaney (Bell Labs) Results (using alt.singles corpus): “As I've commented before, really relating to someone involves standing next to impossible.” “One morning I shot an elephant in my arms and kissed him.” “I spent an interesting evening recently with a grain of salt”

Video Textures Arno Schödl Richard Szeliski David Salesin Irfan Essa Microsoft Research, Georgia Tech

Still photos

Video clips

Video textures

Problem statement video clipvideo texture

Our approach How do we find good transitions?

Finding good transitions Compute L 2 distance D i, j between all frames Similar frames make good transitions frame ivs. frame j

Markov chain representation Similar frames make good transitions

Transition costs Transition from i to j if successor of i is similar to j Cost function: C i  j = D i+1, j

Transition probabilities Probability for transition P i  j inversely related to cost: P i  j ~ exp ( – C i  j /  2 ) high  low 

Preserving dynamics

Cost for transition i  j C i  j = w k D i+k+1, j+k

Preserving dynamics – effect Cost for transition i  j C i  j = w k D i+k+1, j+k

Dead ends No good transition at the end of sequence

Future cost Propagate future transition costs backward Iteratively compute new cost F i  j = C i  j +  min k F j  k

Future cost Propagate future transition costs backward Iteratively compute new cost F i  j = C i  j +  min k F j  k

Future cost Propagate future transition costs backward Iteratively compute new cost F i  j = C i  j +  min k F j  k

Future cost Propagate future transition costs backward Iteratively compute new cost F i  j = C i  j +  min k F j  k

Propagate future transition costs backward Iteratively compute new cost F i  j = C i  j +  min k F j  k Q-learning Future cost

Future cost – effect

Finding good loops Alternative to random transitions Precompute set of loops up front

Visual discontinuities Problem: Visible “Jumps”

Crossfading Solution: Crossfade from one sequence to the other.

Morphing Interpolation task:

Morphing Interpolation task: Compute correspondence between pixels of all frames

Morphing Interpolation task: Compute correspondence between pixels of all frames Interpolate pixel position and color in morphed frame based on [Shum 2000]

Results – crossfading/morphing

Jump Cut Crossfade Morph

Crossfading

Frequent jump & crossfading

Video portrait Useful for web pages

Region-based analysis Divide video up into regions Generate a video texture for each region

Automatic region analysis

Video-based animation Like sprites computer games Extract sprites from real video Interactively control desired motion ©1985 Nintendo of America Inc.

Video sprite extraction

Video sprite control Augmented transition cost:

Interactive fish

Lord of the Flies

Summary Video clips  video textures define Markov process preserve dynamics avoid dead-ends disguise visual discontinuities

Motion Analysis & Synthesis [Efros ’03] What are they doing? –Activity recognition, surveillance, anti-terrorism Can we do the same? –Motion retargeting, movies, video games, etc.

Gathering action data Low resolution, noisy data Moving camera Occlusions

Figure-centric Representation Stabilized spatio-temporal volume –No translation information –All motion caused by person’s limbs Good news: indifferent to camera motion Bad news: hard! Good test to see if actions, not just translation, are being captured

input sequence Remembrance of Things Past “Explain” novel motion sequence with bits and pieces of previously seen video clips Challenge: how to compare motions? motion analysis run walk left swing walk right jog motion synthesis synthetic sequence

How to describe motion? Appearance –Not preserved across different clothing Gradients (spatial, temporal) –same (e.g. contrast reversal) Edges –Too unreliable Optical flow –Explicitly encodes motion –Least affected by appearance –…but too noisy

Motion Descriptor Image frame Optical flow blurred

Comparing motion descriptors t motion-to-motion similarity matrix blurry I … … … … frame-to-frame similarity matrix  I matrix

Recognizing Tennis Red bars show classification results

“Do as I Do” Motion Synthesis Matching two things: –Motion similarity across sequences –Appearance similarity within sequence Dynamic Programming input sequence synthetic sequence

Smoothness for Synthesis is similarity between source and target frames is appearance similarity within target frames For every source frame i, find best target frame by maximizing following cost function: Optimize using dynamic programming

“Do as I Do” Source MotionSource Appearance Result 3400 Frames

“Do as I Say” Synthesis Synthesize given action labels –e.g. video game control run walk left swing walk right jog synthetic sequence run walk left swing walk right jog

“Do as I Say” Red box shows when constraint is applied

Application: Motion Retargeting Rendering new character into existing footage Algorithm –Track original character –Find matches from new character –Erase original character –Render in new character Need to worry about occlusions

Actor Replacement SHOW VIDEO