Presented by Wanxue Dong

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –
Context-based object-class recognition and retrieval by generalized correlograms by J. Amores, N. Sebe and P. Radeva Discussion led by Qi An Duke University.
Salvatore giorgi Ece 8110 machine learning 5/12/2014
Chapter 4: Linear Models for Classification
A Bayesian Approach to Joint Feature Selection and Classifier Design Balaji Krishnapuram, Alexander J. Hartemink, Lawrence Carin, Fellow, IEEE, and Mario.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
The use of unlabeled data to improve supervised learning for text summarization MR Amini, P Gallinari (SIGIR 2002) Slides prepared by Jon Elsas for the.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Expectation Maximization Algorithm
Object Class Recognition Using Discriminative Local Features Gyuri Dorko and Cordelia Schmid.
Presented by Zeehasham Rasheed
Active Appearance Models Computer examples A. Torralba T. F. Cootes, C.J. Taylor, G. J. Edwards M. B. Stegmann.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 20 Learning features one layer at a time Geoffrey Hinton.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Agenda Introduction Bag-of-words models Visual words with spatial location Part-based models Discriminative methods Segmentation and recognition Recognition-based.
Radial-Basis Function Networks
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Crash Course on Machine Learning
Exercise Session 10 – Image Categorization
Presented By Wanchen Lu 2/25/2013
Watch, Listen and Learn Sonal Gupta, Joohyun Kim, Kristen Grauman and Raymond Mooney -Pratiksha Shah.
International Conference on Intelligent and Advanced Systems 2007 Chee-Ming Ting Sh-Hussain Salleh Tian-Swee Tan A. K. Ariff. Jain-De,Lee.
Classifying Images with Visual/Textual Cues By Steven Kappes and Yan Cao.
Unsupervised Learning of Visual Sense Models for Polysemous Words Kate Saenko Trevor Darrell Deepak.
MSRI workshop, January 2005 Object Recognition Collected databases of objects on uniform background (no occlusions, no clutter) Mostly focus on viewpoint.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Object Recognition a Machine Translation Learning a Lexicon for a Fixed Image Vocabulary Miriam Miklofsky.
Sparse Bayesian Learning for Efficient Visual Tracking O. Williams, A. Blake & R. Cipolloa PAMI, Aug Presented by Yuting Qi Machine Learning Reading.
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Computer Vision Lecture 6. Probabilistic Methods in Segmentation.
Topic Models Presented by Iulian Pruteanu Friday, July 28 th, 2006.
Linear Models for Classification
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
 Effective Multi-Label Active Learning for Text Classification Bishan yang, Juan-Tao Sun, Tengjiao Wang, Zheng Chen KDD’ 09 Supervisor: Koh Jia-Ling Presenter:
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe.
Computational Intelligence: Methods and Applications Lecture 26 Density estimation, Expectation Maximization. Włodzisław Duch Dept. of Informatics, UMK.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
A Self-organizing Semantic Map for Information Retrieval Xia Lin, Dagobert Soergel, Gary Marchionini presented by Yi-Ting.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Recent Trends in Text Mining
Deep Feedforward Networks
CSC321: Neural Networks Lecture 22 Learning features one layer at a time Geoffrey Hinton.
University of Ioannina
Adversarial Learning for Neural Dialogue Generation
Restricted Boltzmann Machines for Classification
Classification of unlabeled data:
Computer vision: models, learning and inference
Multimodal Learning with Deep Boltzmann Machines
10701 / Machine Learning.
Introductory Seminar on Research: Fall 2017
Classification with Perceptrons Reading:
Machine Learning Basics
Machine Learning Ali Ghodsi Department of Statistics
Latent Variables, Mixture Models and EM
Unsupervised Learning and Autoencoders
“Bayesian Identity Clustering”
Learning with information of features
10701 / Machine Learning Today: - Cross validation,
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
The Naïve Bayes (NB) Classifier
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
EM Algorithm and its Applications
Unsupervised learning of visual sense models for Polysemous words
Presentation transcript:

Presented by Wanxue Dong Who’s in the Picture Presented by Wanxue Dong

Research Question: Who is depicted in the associated image? Find a correspondence between the faces and names. Why this is important? It has been used to browse museum collections and organize large image collections To take the large collection of news images and captions as semi- supervised input and produce a fully supervised dataset of faces labeled with names

Linking a face and language model with EM (Expectation Maximization) Consider name assignment as a hidden variable problem where the hidden variables are the correct name-face correspondences for each picture. EM (Expectation Maximization) procedure: iterates between computing the expected values of the set of face-name correspondences and updating the face clusters and language model given the correspondences.

Generative Model

Name Assignment The likelihood of picture xi, under assignment ai, of names to faces under our generative model is α indexes into the names that are pictured σ(α) indexes into the faces assigned to the pictured names β indexes into the names that are not pictured γ indexes into the faces without assigned names

Name Assignment EM procedure: E – update the Pij according to the normalized probability of picture i with assignment j. M – Maximize the parameters P(face | name) and P(pictured | context)

Modeling the Appearance of Faces P(face | name) P(face | name) gaussians with fixed covariance Need a representation for faces in a feature space Rectify all faces to a canonical pose Five support vector machines to train the feature detectors Use kPCA to reduce the dimensionality of data, compute linear discriminants

Language Model P(pictured | context) Assigns a probability to each name based on its context within the caption. The distributions learned using counts of how often each context appears describing an assigned name and unassigned name One distribution for each context cue modeled independently

Comparison of EM & MM procedure The Maximal Assignment procedure: M1 – set of maximal Pij to 1 and all others to 0 M2 – maximize the parameters P(face | name) and P(pictured | context) For both methods, incorporating a language model improves their respective clustering greatly.

Data Sets Consisting of approximately half a million news pictures and captions from Yahoo News over a period of roughly two years Faces: 44,773 large well detected face images Names: use an open source named entity recognizer to detect proper names Scale: reject face images that cannot be rectified satisfactorily, leaving 34,623. Finally, concentrate on images within whose captions could be detected proper names, leaving 30,281.

Experiments’ Results Results of applying learned language model to a test set of 430 captions (text alone) Test set: hand labeled each detected name with IN/OUT based on whether the referred name was pictured within the corresponding picture Test how well the language model could predict those labels

Conclusion This study coupled language and images, using language to learn about images and images to learn about language Analyzing language more carefully can produce a much better clustering A natural language classifier that can be used to determine who is pictured from text alone

Critiques and Future Work Test set is limited as all hand labeled There is no other model comparison Next step: learn a language model for free text on a webpage to improve google image search results

Reference Tamara L. Berg, Alexander C. Berg, Jaety Edwards, David A. Forsyth, Who's in the Picture?, Neural Information Processing Systems (NIPS), 2004. Thank you!