Colorization by Example R. Irony, D. Cohen-Or, D. Lischinski Tel-Aviv University The Hebrew University of Jerusalem Eurgraphics Symposium on Rendering.

Slides:



Advertisements
Similar presentations
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Advertisements

Medical Image Registration Kumar Rajamani. Registration Spatial transform that maps points from one image to corresponding points in another image.
A CTION R ECOGNITION FROM V IDEO U SING F EATURE C OVARIANCE M ATRICES Kai Guo, Prakash Ishwar, Senior Member, IEEE, and Janusz Konrad, Fellow, IEEE.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Automatic Feature Extraction for Multi-view 3D Face Recognition
Proportion Priors for Image Sequence Segmentation Claudia Nieuwenhuis, etc. ICCV 2013 Oral.
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
September 10, 2013Computer Vision Lecture 3: Binary Image Processing 1Thresholding Here, the right image is created from the left image by thresholding,
Color2Gray: Salience-Preserving Color Removal Amy A. Gooch Sven C. Olsen Jack Tumblin Bruce Gooch.
Watching Unlabeled Video Helps Learn New Human Actions from Very Few Labeled Snapshots Chao-Yeh Chen and Kristen Grauman University of Texas at Austin.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean 5403 T-R 3:00pm – 4:20pm Lecture #20.
A new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs Combines both geometry-based and image.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Abstract Extracting a matte by previous approaches require the input image to be pre-segmented into three regions (trimap). This pre-segmentation based.
KNN, LVQ, SOM. Instance Based Learning K-Nearest Neighbor Algorithm (LVQ) Learning Vector Quantization (SOM) Self Organizing Maps.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
Image Analogies Aaron Hertzmann (1,2) Charles E. Jacobs (2) Nuria Oliver (2) Brian Curless (3) David H. Salesin (2,3) 1 New York University 1 New York.
Ron Yanovich & Guy Peled 1. Contents Grayscale coloring background Luminance / Luminance channel Segmentation Discrete Cosine Transform K-nearest-neighbor.
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
New Segmentation Methods Advisor : 丁建均 Jian-Jiun Ding Presenter : 蔡佳豪 Chia-Hao Tsai Date: Digital Image and Signal Processing Lab Graduate Institute.
Prakash Chockalingam Clemson University Non-Rigid Multi-Modal Object Tracking Using Gaussian Mixture Models Committee Members Dr Stan Birchfield (chair)
Presented by Tienwei Tsai July, 2005
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
Ajay Kumar, Member, IEEE, and David Zhang, Senior Member, IEEE.
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
Digital Image Processing CCS331 Relationships of Pixel 1.
Global Parametrization of Range Image Sets Nico Pietroni, Marco Tarini, Olga Sorkine, Denis Zorin.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Pattern Recognition April 19, 2007 Suggested Reading: Horn Chapter 14.
Face Recognition: An Introduction
CSE 185 Introduction to Computer Vision Face Recognition.
2005/12/021 Content-Based Image Retrieval Using Grey Relational Analysis Dept. of Computer Engineering Tatung University Presenter: Tienwei Tsai ( 蔡殿偉.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
Fast Query-Optimized Kernel Machine Classification Via Incremental Approximate Nearest Support Vectors by Dennis DeCoste and Dominic Mazzoni International.
Dimensionality reduction
Nottingham Image Analysis School, 23 – 25 June NITS Image Segmentation Guoping Qiu School of Computer Science, University of Nottingham
Computational Biology Group. Class prediction of tumor samples Supervised Clustering Detection of Subgroups in a Class.
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
2D-LDA: A statistical linear discriminant analysis for image matrix
Instructor: Mircea Nicolescu Lecture 5 CS 485 / 685 Computer Vision.
Color Image Segmentation Mentor : Dr. Rajeev Srivastava Students: Achit Kumar Ojha Aseem Kumar Akshay Tyagi.
Unsupervised Classification
Shadow Detection in Remotely Sensed Images Based on Self-Adaptive Feature Selection Jiahang Liu, Tao Fang, and Deren Li IEEE TRANSACTIONS ON GEOSCIENCE.
Shape2Pose: Human Centric Shape Analysis CMPT888 Vladimir G. Kim Siddhartha Chaudhuri Leonidas Guibas Thomas Funkhouser Stanford University Princeton University.
A Self-organizing Semantic Map for Information Retrieval Xia Lin, Dagobert Soergel, Gary Marchionini presented by Yi-Ting.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
图像处理技术讲座(3) Digital Image Processing (3) Basic Image Operations
Color Image Processing
IMAGE PROCESSING RECOGNITION AND CLASSIFICATION
Color Image Processing
Mean Shift Segmentation
Computer Vision Lecture 5: Binary Image Processing
Color Image Processing
Color Image Processing
Feature space tansformation methods
Presentation transcript:

Colorization by Example R. Irony, D. Cohen-Or, D. Lischinski Tel-Aviv University The Hebrew University of Jerusalem Eurgraphics Symposium on Rendering (2005)

Introduction Levin et al. [LLW04] recently proposed a simple yet effective user-guided colorization method. In this method the user is required to scribble the desired colors in the interiors of the various regions. Automatically propagates the scribbled colors to produce a completely colorized image.

Introduction Levin et al.’s colorization. Left: Given a grayscale image marked with some color scribbles by the user. Middle: Levin et al’s colorization produces a colorized image. Right: For reference, the original color image is shown.

Introduction Levin et al.’s colorization. Left: dozens of user drawn scribbles (some very small). Right: resulting colorization. This method might require the user to carefully place a multitude of appropriately colored scribbles.

Introduction Welsh et al. [WAM02] proposed an automatic colorization technique that colorizes an image by matching small pixel neighborhoods in the image to those in the reference image, and transferring colors accordingly.

Supervised Classification Supervised classification methods typically consist of two phases: feature analysis classification In this paper we adopt a classification approach based on the K-nearest-neighbor (Knn) rule [DHS00]. This is an extremely simple yet effective method for supervised classification.

Supervised Classification Linear dimensionality reduction techniques are often used to make such classifiers both more efficient and more effective. For example, PCA-based techniques apply a linear projection that reduces the dimension of the data while maximizing the scatter of all projected samples. Linear Discriminant Analysis (LDA) [BHK97, DHS00, Fis36], which finds a linear subspace in which the ratio of between-class scatter to that of within-class scatter is maximized.

Our approach vs. color transfer.

Colorization by Example Our algorithm colorizes one or more input grayscale images, based on a partially segmented reference color image. By partial segmentation we mean that one or more mutually disjoint regions in the image and each region has been assigned a unique label. Such segmentations may be either computed automatically, or marked manually by the user.

Colorization by Example This approach consists of the following main conceptual stages: Training Classification Color transfer Optimization

Overview of our colorization technique

Training The luminance channel of the reference image along with the accompanying partial segmentation are provided as a training set to supervised learning algorithm. This algorithm constructs a low-dimensional feature space in which it is easy to discriminate between pixels belonging to differently labeled regions.

Feature Spaces and Classifier Given the reference color image and its partial segmentation, our first task is to construct a feature space and a corresponding classifier. Every pixel in one of these regions defines a labeled feature vector, a point in the feature space. Given a previously unseen feature vector, the goal of the classifier is to decide which label should be assigned to it.

Feature Spaces and Classifier The classifier must be able to distinguish between different classes mainly based on texture. We should associate each pixel with a feature vector representing its monochromatic texture. We use the Discrete Cosine Transform (DCT) coefficients of a k by k neighborhood around the pixel as its feature vector.

Feature Spaces and Classifier The DCT transform yields a k 2 -dimensional feature space, populated with the labeled feature vectors corresponding to the training set pixels. A novel feature vector may be naively classified by assigning it the label of its nearest feature space neighbor. A sophisticated classifier is defined by the K- nearest-neighbor (Knn) rule.

Feature Spaces and Classifier This classifier examines the K nearest neighbors of the feature vector and chooses the label by a majority vote. Better results may be obtained by switching to a low-dimensional subspace, custom- tailored according to the training set, using an approach similar to linear discriminant analysis (LDA).

Applying Knn in a discriminating subspace: The feature space is populated by points belonging to two classes: magenta and cyan. The yellow highlighted point has a majority of magenta- colored nearest neighbors. After rotating the space to the UV coordinate system V is the principle direction of the intra-difference vectors, and then projecting the points onto the U axis, all of the nearest neighbors are cyan.

Simple Knn-matching based on similar luminance value and neighborhood statistics (e) vs. our matching (g).

Feature Spaces and Classifier A transformation T which transforms the vector of k 2 DCT coefficients to a point in the low-dimensional subspace. The distance between pixels p and q : f (x) is the vector of DCT coefficients corresponding to the k×k neighborhood centered at x.

Although the Knn classifier described above is more robust than a naive nearest-neighbor classifier, there can still be many misclassified pixels. Quite a few pixels inside the body of the cheetah are classified as belonging to the background, and vice versa.

Image space Voting N(p), the k×k neighborhood around a pixel p in the input image. We replace the label of p with the dominant label in N(p). The dominant label is the label with the highest confidence.

Image space Voting Wq depend on the distance D(q,Mq), between the pixel q and its best match Mq. Mq is the nearest neighbor of q in the feature space, which has the same label as q. is the set of pixels in N(p) with the label

(c) Shows how this image space voting improves the spatial coherence of the resulting classification

Color Transfer and Optimization We work in the YUV color space, where Y is the monochromatic luminance channel, which we use to perform the classification, while U and V are the chrominance channels. C(p) denote the chrominance coordinates of a pixel p. The color of p (with label ) is given by the weighted average C(p)

Color Transfer and Optimization Mq(p) denotes the pixel in example image whose position with respect to Mq is the same as the position of p with respect to q. Assigning color to pixel p: each neighbor of p (e.g., q, r) has a matching neighborhood in the reference image, which “predicts” a different color for p. The color of p is a weighted average of these predictions.

We transfer color only to pixels whose confidence in their label is sufficiently large, aaa > 0.5, and provide the colored pixels as constraints to the optimization-based color interpolation scheme of Levin et al. [LLW04]. The optimization-based interpolation is based on the principle that neighboring pixels with similar luminance should have similar colors. Color Transfer and Optimization

The interpolation attempts to minimize the difference between the color assigned to a pixel p and the weighted average of the colors of its neighbors. Color Transfer and Optimization W pq is a weighting function that sums to one

Generating automatic scribbles: pixels with confidence above a predefined threshold are provided as input to the optimization stage.

Closeup of the head: (a)before optimization (b)confidence map (c)final result after optimization.

Result We compare our results with the results achieved by automatic pixelwise color transfer [WAM02]. We further show that naive classifiers alone do not yield satisfactory results and that our feature space analysis and image space voting greatly improve classification quality.

Result

We also compare our method to the user- assisted method of Levin et al. [LLW04] The advantage of their technique is that it does not require a reference image. It is sometimes sufficient to manually colorize a small part of the image using Levin’s method, and then complete the colorization automatically, using that small part as the reference.

Result A significant advantage of our method lies in its ability to automatically colorize an entire series of grayscale images.

Result Colorization of a series of grayscale images from a single example. Left column: the reference image, its collection of regions, and its luminance channel. On the right: Top row : the input grayscale images; Middle row : corresponding classifications; Bottom row : the resulting colorizations.

Future Work In the future we plan to investigate the possibility of combining our classification-based approach with automatically established swatches. This may be done by searching for representative regions in the input images that match reference regions particularly well. We would also like to explore other, more sophisticated, monochrome texture descriptors.