SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Learning deformable models Yali Amit, University of Chicago Alain Trouvé, CMLA Cachan.
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
A Bayesian Approach to Recognition Moshe Blank Ita Lifshitz Reverend Thomas Bayes
Pattern Recognition and Machine Learning
Clustering with k-means and mixture of Gaussian densities Jakob Verbeek December 3, 2010 Course website:
Computer Vision for Human-Computer InteractionResearch Group, Universität Karlsruhe (TH) cv:hci Dr. Edgar Seemann 1 Computer Vision: Histograms of Oriented.
Digital Image Processing In The Name Of God Digital Image Processing Lecture3: Image enhancement M. Ghelich Oghli By: M. Ghelich Oghli
AdaBoost & Its Applications
Face detection Many slides adapted from P. Viola.
What is Statistical Modeling
Visual Recognition Tutorial
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Classification and risk prediction
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Generic Object Recognition -- by Yatharth Saraf A Project on.
Unsupervised Learning: Clustering Rong Jin Outline  Unsupervised learning  K means for clustering  Expectation Maximization algorithm for clustering.
Lecture 5: Learning models using EM
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Expectation Maximization for GMM Comp344 Tutorial Kai Zhang.
Visual Recognition Tutorial
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Crash Course on Machine Learning
Computer vision: models, learning and inference Chapter 6 Learning and Inference in Vision.
Face Detection using the Viola-Jones Method
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Building local part models for category-level recognition C. Schmid, INRIA Grenoble Joint work with G. Dorko, S. Lazebnik, J. Ponce.
REVISED CONTEXTUAL LRT FOR VOICE ACTIVITY DETECTION Javier Ram’ırez, Jos’e C. Segura and J.M. G’orriz Dept. of Signal Theory Networking and Communications.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Window-based models for generic object detection Mei-Chen Yeh 04/24/2012.
Object Detection with Discriminatively Trained Part Based Models
Baseband Demodulation/Detection
Automatic Image Annotation by Using Concept-Sensitive Salient Objects for Image Content Representation Jianping Fan, Yuli Gao, Hangzai Luo, Guangyou Xu.
Bayesian Parameter Estimation Liad Serruya. Agenda Introduction Bayesian decision theory Scale-Invariant Learning Bayesian “One-Shot” Learning.
Chapter 4: Pattern Recognition. Classification is a process that assigns a label to an object according to some representation of the object’s properties.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Fields of Experts: A Framework for Learning Image Priors (Mon) Young Ki Baik, Computer Vision Lab.
Face Detection Ying Wu Electrical and Computer Engineering Northwestern University, Evanston, IL
Expectation-Maximization (EM) Case Studies
Robust Real Time Face Detection
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Paper Reading Dalong Du Nov.27, Papers Leon Gu and Takeo Kanade. A Generative Shape Regularization Model for Robust Face Alignment. ECCV08. Yan.
Boosted Particle Filter: Multitarget Detection and Tracking Fayin Li.
Prototype Classification Methods Fu Chang Institute of Information Science Academia Sinica ext. 1819
Lecture 2: Statistical learning primer for biologists
Flat clustering approaches
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Probability and Statistics in Vision. Probability Objects not all the sameObjects not all the same – Many possible shapes for people, cars, … – Skin has.
October 1, 2013Computer Vision Lecture 9: From Edges to Contours 1 Canny Edge Detector However, usually there will still be noise in the array E[i, j],
Lecture 5: Statistical Methods for Classification CAP 5415: Computer Vision Fall 2006.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Meeting 8: Features for Object Classification Ullman et al.
Zhaoxia Fu, Yan Han Measurement Volume 45, Issue 4, May 2012, Pages 650–655 Reporter: Jing-Siang, Chen.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Unsupervised Learning Part 2. Topics How to determine the K in K-means? Hierarchical clustering Soft clustering with Gaussian mixture models Expectation-Maximization.
Lecture 1.31 Criteria for optimal reception of radio signals.
Computer vision: models, learning and inference
Fast Kernel-Density-Based Classification and Clustering Using P-Trees
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Classification of unlabeled data:
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Image Segmentation Techniques
Probabilistic Models with Latent Variables
Pattern Recognition and Machine Learning
EM Algorithm and its Applications
Presentation transcript:

SVCL Automatic detection of object based Region-of-Interest for image compression Sunhyoung Han

SVCL Transmission in erroneous channel Basic Motivation Spatially different Super resolution Constraints Limited Resources & Channel Errors

SVCL Basic motivation By having information about importance of regions One can wisely use the limited resources

SVCL User-adaptive Coder v visual concepts of interest can be anything main idea: let users define a universe of objects of interest train saliency detector for each object e.g. regions of “people”, “the Capitol”, “trees”, etc.

SVCL User Adaptive Coder query provided by user train detector current training sets

SVCL User-adaptive coder user-adaptive coder: –detector should be generic enough to handle large numbers of object categories –training needs to be reasonably fast (including example preparation time) “face”“lamp”“car”

SVCL User-adaptive coder proposed detector –top-down object detector (object category specified by user) –focus on weak supervision instead of highly accurate localization –composed of saliency detection and saliency validation –discriminant saliency: saliency filters training FIND best features

SVCL Discriminant Saliency start from a universe of classes (e.g. “faces”, “trees”, “cars”, etc.) design a dictionary of features: e.g. linear combinations of DCT coefficients at multiple scales salient features: those that best distinguish the object class of interest from random background scenes. salient regions are the regions of the image where these detectors have strong response see [Gao & Vasconcelos, NIPS, 2004].

SVCL Top-down Discriminant Saliency Model Scale Selection W j WTA Faces Discriminant Feature Selection Salient Features Background Saliency Map Original Feature Set Malik-Perona pre-attentive perception model

SVCL saliency detector salient point sal i : –magnitude  i –location l i –scale s i saliency map approximated by a Gaussian mixture Saliency representation image saliency mapsalient points Probability map

SVCL Saliency validation saliency detection: –due to limited feature dictionary and/or limited training set –coarse detection of object class of interest need to eliminate false positives saliency validation: –geometric consistency –reject salient points whose spatial configuration is inconsistent with training examples original Image saliency map for ‘street sign’ example of saliency map

SVCL Saliency validation learning a geometric model of salient point con- figuration two components : - image alignment model: - classify points into true positives - configuration model false positives - model each as Gaussian

SVCL Saliency validation model: two classes of points Y={0,1} –Y=1 true positive –Y=0 false positive saliency map: mixture of true and false positive saliency distributions each distribution approximated by a Gaussian

SVCL this is a two class clustering problem –can be solved by expectation-maximization graphical model non-standard issues –we start from distributions, not points –alignment does not depend on false negatives Saliency validation E-stepM-step YX  L~uniform Y~Bernoulli (  1 ) C| Y=i ~multinomial (  i ) X| Y=i,L=l,S=s,  ~G(x, l- ,  ) L,S C

SVCL Saliency Validation For K training examples (# of saliency point is N k for kth example) Missing data  Y= j, j ∈ {1,0} Parameters   j (probability for class j) ∑ j (Covariance for class )  k (displacement for kth example) For robust update DERIVATION DETAILS

SVCL Saliency Validation visualization of EM algorithm Saliency detection result Init saliency points overlapped over 40 samples Visualized variance ∑ 1 Overlapped points classified as ‘’object’’ Overlapped points classified as ‘’noise’’ Visualized variance ∑ 0

SVCL Saliency Validation examples of classified Points in summary, during training we learn –discriminant features –The “right” configuration of salient points Examples of classified saliency points  White if h ij 1 >h ij 0 Black otherwise

SVCL Region of interest detection find image window that best matches the learned configuration mathematically: - find location p where the posterior probability of the object class is the largest

SVCL Region of interest detection by Bayes rule –Posterior  Likelihood x Prior –likelihood is given by matching saliencies within the window & the model - prior measures the saliency mass inside window ? ? likelihood Prior

SVCL Region of Interest Detection given the model –the likelihood, under it, of a set of points drawn from the observed saliency distribution is –and the optimal location is given by Prior for location P With saliency detector DERIVATION DETAILS Measure configuration matching

SVCL 2. Determine scale(shape) of ROI mask Observation(∑ * ) from data and prior(∑ 1 ) from training data are used 3. Thresholds P Y|X,P (1|x,p*) to get binary ROI mask Region of Interest Detection ** Once the center point is known the assignment of each point is given by  The observed configuration for Y=1 is x ∑1∑1 ∑*∑*

SVCL Region of Interest Detection Saliency detection (for statue of liberty) Probability map (saliency only) Probability map (with configuration info.) ROI mask Example of ROI Detection

SVCL Evaluation Using CalTech “Face” database & UIUC “Car side” database Evaluate robustness of learning –Dedicated Training set vs. Web Training set Evaluation Metric –ROC area curve –PSNR gain for ROI coding vs. normal coding Number of positive example: 550 Number of positive example: 100

SVCL Evaluation ROC area curve False Positive True Positive False Positive True Positive “Car” “Face”

SVCL Evaluation PSNR performance comparison “Car”“Face” Bit Per Pixel PSNR Bit Per Pixel PSNR 14.3% bits can be saved even with web train uniform case for the same image quality

SVCL Result Examples

SVCL Result Comparison of needed bits to get the same PSNR (30 dB) for ROI  Maximally, ¼ bits are enough to get the same quality for ROI area

SVCL Result Examples Normal coding ROI coding

SVCL

EM derivation Want to fit lower level observation For a virtual sample X = {X ik |i=1, …, N k and k=1, …, K} with the size of M ik = ik *N, likelihood becomes For complete set the log likelihood becomes 

SVCL EM derivation  Maximization in the m-step is carried out by maximizing the Lagrangian

SVCL ROI Detection For one sample point x 1  For samples having distribution of

SVCL ROI Detection Therefore,