A Two-Step Approach to Hallucinating Faces: Global Parametric Model and Local Nonparametric Model Ce Liu Heung-Yeung Shum Chang Shui Zhang CVPR 2001.

Slides:



Advertisements
Similar presentations
Part 2: Unsupervised Learning
Advertisements

Basic Steps 1.Compute the x and y image derivatives 2.Classify each derivative as being caused by either shading or a reflectance change 3.Set derivatives.
Example 1 Generating Random Photorealistic Objects Umar Mohammed and Simon Prince Department of Computer Science, University.
Removing blur due to camera shake from images. William T. Freeman Joint work with Rob Fergus, Anat Levin, Yair Weiss, Fredo Durand, Aaron Hertzman, Sam.
Bayesian Belief Propagation
Image Super-resolution via Sparse Representation
Bayesian Reconstruction of 3D Human Motion from Single-Camera Video
Optimizing and Learning for Super-resolution
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
Various Regularization Methods in Computer Vision Min-Gyu Park Computer Vision Lab. School of Information and Communications GIST.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Hongliang Li, Senior Member, IEEE, Linfeng Xu, Member, IEEE, and Guanghui Liu Face Hallucination via Similarity Constraints.
An Overview of Machine Learning
Learning Inhomogeneous Gibbs Models Ce Liu
Learning to estimate human pose with data driven belief propagation Gang Hua, Ming-Hsuan Yang, Ying Wu CVPR 05.
Human Pose detection Abhinav Golas S. Arun Nair. Overview Problem Previous solutions Solution, details.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
GrabCut Interactive Image (and Stereo) Segmentation Carsten Rother Vladimir Kolmogorov Andrew Blake Antonio Criminisi Geoffrey Cross [based on Siggraph.
IMAGE RESTORATION AND REALISM MILLIONS OF IMAGES SEMINAR YUVAL RADO.
1 Removing Camera Shake from a Single Photograph Rob Fergus, Barun Singh, Aaron Hertzmann, Sam T. Roweis and William T. Freeman ACM SIGGRAPH 2006, Boston,
Rob Fergus Courant Institute of Mathematical Sciences New York University A Variational Approach to Blind Image Deconvolution.
Optimization & Learning for Registration of Moving Dynamic Textures Junzhou Huang 1, Xiaolei Huang 2, Dimitris Metaxas 1 Rutgers University 1, Lehigh University.
Personal Photo Enhancement using Example Images Neel Joshi Wojciech Matusik, Edward H. Adelson, and David J. Kriegman Microsoft Research, Disney Research,
Face Verification across Age Progression Narayanan Ramanathan Dr. Rama Chellappa.
Announcements Project 4 questions? Guest lectures Thursday: Richard Ladner “tactile graphics” Next Tuesday: Jenny Yuen and Jeff Bigham.
Face Poser: Interactive Modeling of 3D Facial Expressions Using Model Priors Manfred Lau 1,3 Jinxiang Chai 2 Ying-Qing Xu 3 Heung-Yeung Shum 3 1 Carnegie.
Texture Splicing Yiming Liu, Jiaping Wang, Su Xue, Xin Tong, Sing Bing Kang, Baining Guo.
Curve Analogies Aaron Hertzmann Nuria Oliver Brain Curless Steven M. Seitz University of Washington Microsoft Research Thirteenth Eurographics.
Learning Low-Level Vision William T. Freeman Egon C. Pasztor Owen T. Carmichael.
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
An Iterative Optimization Approach for Unified Image Segmentation and Matting Hello everyone, my name is Jue Wang, I’m glad to be here to present our paper.
Radial Basis Function Networks
Noise Estimation from a Single Image Ce Liu William T. FreemanRichard Szeliski Sing Bing Kang.
Super-Resolution of Remotely-Sensed Images Using a Learning-Based Approach Isabelle Bégin and Frank P. Ferrie Abstract Super-resolution addresses the problem.
3D-Assisted Facial Texture Super-Resolution Pouria Mortazavian, Josef Kittler, William Christmas 10 September 2009 Centre for Vision, Speech and Signal.
Image Analogies Aaron Hertzmann (1,2) Charles E. Jacobs (2) Nuria Oliver (2) Brian Curless (3) David H. Salesin (2,3) 1 New York University 1 New York.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Step 3: Classification Learn a decision rule (classifier) assigning bag-of-features representations of images to different classes Decision boundary Zebra.
Last tuesday, you talked about active shape models Data set of 1,500 hand-labeled faces 20 facial features (eyes, eye brows, nose, mouth, chin) Train 40.
City University of Hong Kong 18 th Intl. Conf. Pattern Recognition Self-Validated and Spatially Coherent Clustering with NS-MRF and Graph Cuts Wei Feng.
Object Stereo- Joint Stereo Matching and Object Segmentation Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on Michael Bleyer Vienna.
Texture Optimization for Example-based Synthesis Vivek Kwatra Irfan Essa Aaron Bobick Nipun Kwatra.
Markov Random Fields Probabilistic Models for Images
Structured Face Hallucination Chih-Yuan Yang Sifei Liu Ming-Hsuan Yang Electrical Engineering and Computer Science 1.
The 18th Meeting on Image Recognition and Understanding 2015/7/29 Depth Image Enhancement Using Local Tangent Plane Approximations Kiyoshi MatsuoYoshimitsu.
Paper Reading Dalong Du Nov.27, Papers Leon Gu and Takeo Kanade. A Generative Shape Regularization Model for Robust Face Alignment. ECCV08. Yan.
3D Face Recognition Using Range Images
Lecture 2: Statistical learning primer for biologists
AAM based Face Tracking with Temporal Matching and Face Segmentation Mingcai Zhou 1 、 Lin Liang 2 、 Jian Sun 2 、 Yangsheng Wang 1 1 Institute of Automation.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Dynamic Background Learning through Deep Auto-encoder Networks Pei Xu 1, Mao Ye 1, Xue Li 2, Qihe Liu 1, Yi Yang 2 and Jian Ding 3 1.University of Electronic.
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
SIGGRAPH 2007 Hui Fang and John C. Hart.  We propose an image editing system ◦ Preserve its detail and orientation by resynthesizing texture from the.
ICCV 2007 Optimization & Learning for Registration of Moving Dynamic Textures Junzhou Huang 1, Xiaolei Huang 2, Dimitris Metaxas 1 Rutgers University 1,
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Edge Preserving Spatially Varying Mixtures for Image Segmentation Giorgos Sfikas, Christophoros Nikou, Nikolaos Galatsanos (CVPR 2008) Presented by Lihan.
Biointelligence Laboratory, Seoul National University
Announcements Project 4 out today help session at the end of class.
Nonparametric Semantic Segmentation
Synthesis of X-ray Projections via Deep Learning
Probabilistic Models for Linear Regression
Data-driven methods: Texture 2 (Sz 10.5)
Binarization of Low Quality Text Using a Markov Random Field Model
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Announcements Guest lecture next Tuesday
Image and Video Processing
LECTURE 07: BAYESIAN ESTIMATION
Presentation transcript:

A Two-Step Approach to Hallucinating Faces: Global Parametric Model and Local Nonparametric Model Ce Liu Heung-Yeung Shum Chang Shui Zhang CVPR 2001

Face hallucination Face Hallucination to infer high resolution face image from low resolution input (a) Input 24×32 (b) Hallucinated result (c) Original 96×128

Why to study face hallucination? Applications Video conference To use very low band to transmit face image sequence To repair damaged images in transmission Face image recovery To recover low-quality faces in old photos To recover low-resolution monitoring videos Research Information recovery How to formulate and learn prior knowledge of face How to apply face prior to infer the lost high frequency details Super resolution How to model the bridge from low-resolution to high-resolution

Difficulties and solution strategy Difficulties Sanity Constraint The result must be close to the input image when smoothed and down-sampled Global Constraint The result must have common characteristics of a human face, e.g., eyes, mouth, nose and symmetry Local Constraint The result must have specific characteristics of this face image, with photorealistic local features Solution strategy We choose learning based method aided by a large set of various face images to hallucinate face

Previous learning-based super-resolution methods Multi-resolution texture synthesis De Bonet. SIGGRAPH 1997 Markov network Freeman and Pasztor. ICCV 1999 Face hallucination Baker and Kanade. AFGR 2000, CVPR 2000 Image analogies Hertzmann, Jacobs, Oliver, Curless and Salesin. SIGGRAPH 2001 They all use local feature transfer or inference in Markov random field, without any global correspondence taken into account.

Decouple high-resolution face image to two parts high resolution face image global face local face Two-step Bayesian inference 1. Inferring global face 2. Inferring local face Finally adding them together Our method 1. Inferring global face 2. Inferring local face Finally adding them together ?

Flowchart of hallucinating face Learning Process Inference Process Training dataset Global faces Local faces Learning (a) Learn the prior of global face by PCA (b) Build Markov network between global and local faces Inference (c) Infer global face by linear regression (d) Infer local face by Markov network (c) (d) (a) (b) InputOutput

Inferring global face Prior Assume the prior of global face to be Gaussian and learn it by PCA. The global face is the principal components of the high-resolution face image. (Many other methods such as Gaussian mixture, ICA, kernel PCA, TCA can be used to model the face prior. We choose PCA because it could get simple solution) Likelihood Treat low resolution input as a soft constraint to the global face. The likelihood turns out to be a Gaussian distribution again. Posteriori The energy of the posteriori has a quadratic form. The MAP solution is converted to linear regression by SVD.

How to compute global face

Local face is pursued by minimizing the energy of Markov network Two terms of energies: external potential to model the connective statistics between two linked patches in and. internal potential to make adjacent patches in well connected. Energy minimization by simulated annealing Inferring local face by Markov network An inhomogeneous patch-based nonparametric Markov network

Experimental results (1) (a)(b)(c)(d) (a)Input low 24×32 (b)Inferred global face (c)Hallucinated result (d)Original high 96×128

Experimental results (2)

Experimental results (3)

Comparison with other methods (a)(b)(c)(d)(e)(f) (a)Input (b)Hallucinated by our method (c)Cubic B-spline (d)Hertzmann et al. (e)Baker et al. (f)Original

Summary Hybrid modeling of face (global plus local) Global: the major information of face, lying in middle and low frequency band Local: the residue between real data and global model, lying in high frequency band The sanity constraint is added to the global part The global face is modeled by PCA and inferred by linear regression The conditional distribution of the local face given the global face is modeled upon a patch-based nonparametric Markov network, and inferred by energy minimization Both of the two steps in inference are global optimal Global part: optimizing a quadratic energy function by SVD Local part: optimizing the network energy by simulated annealing

Thank you!