Discriminative Probabilistic Models for Relational Data

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Learning on the Test Data: Leveraging “Unseen” Features Ben Taskar Ming FaiWong Daphne Koller.
Graphical Models BRML Chapter 4 1. the zoo of graphical models Markov networks Belief networks Chain graphs (Belief and Markov ) Factor graphs =>they.
Constrained Approximate Maximum Entropy Learning (CAMEL) Varun Ganapathi, David Vickrey, John Duchi, Daphne Koller Stanford University TexPoint fonts used.
Maximum Margin Markov Network Ben Taskar, Carlos Guestrin Daphne Koller 2004.
Parameter Learning in MN. Outline CRF Learning CRF for 2-d image segmentation IPF parameter sharing revisited.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
A Statistical Model for Domain- Independent Text Segmentation Masao Utiyama and Hitoshi Isahura Presentation by Matthew Waymost.
Relational Learning with Gaussian Processes By Wei Chu, Vikas Sindhwani, Zoubin Ghahramani, S.Sathiya Keerthi (Columbia, Chicago, Cambridge, Yahoo!) Presented.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Carnegie Mellon Evidence-Specific Structures for Rich Tractable CRFs Anton Chechetka, Carlos Guestrin General approach: 1.Ground the model / features 2.Use.
MSRC Summer School - 30/06/2009 Cambridge – UK Hybrids of generative and discriminative methods for machine learning.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
Oregon State University – CS539 PRMs Learning Probabilistic Models of Link Structure Getoor, Friedman, Koller, Taskar.
Co-training LING 572 Fei Xia 02/21/06. Overview Proposed by Blum and Mitchell (1998) Important work: –(Nigam and Ghani, 2000) –(Goldman and Zhou, 2000)
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Bayesian Networks Alan Ritter.
11 CS 388: Natural Language Processing: Discriminative Training and Conditional Random Fields (CRFs) for Sequence Labeling Raymond J. Mooney University.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Cao et al. ICML 2010 Presented by Danushka Bollegala.
Modeling Relationship Strength in Online Social Networks Rongjing Xiang: Purdue University Jennifer Neville: Purdue University Monica Rogati: LinkedIn.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
Sum-Product Networks CS886 Topics in Natural Language Processing
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
1 CS 391L: Machine Learning: Bayesian Learning: Beyond Naïve Bayes Raymond J. Mooney University of Texas at Austin.
Collective Classification A brief overview and possible connections to -acts classification Vitor R. Carvalho Text Learning Group Meetings, Carnegie.
Aprendizagem Computacional Gladys Castillo, UA Bayesian Networks Classifiers Gladys Castillo University of Aveiro.
1 Generative and Discriminative Models Jie Tang Department of Computer Science & Technology Tsinghua University 2012.
Markov Random Fields Probabilistic Models for Images
An Asymptotic Analysis of Generative, Discriminative, and Pseudolikelihood Estimators by Percy Liang and Michael Jordan (ICML 2008 ) Presented by Lihan.
Max-Margin Markov Networks by Ben Taskar, Carlos Guestrin, and Daphne Koller Presented by Michael Cafarella CSE574 May 25, 2005.
Graph-based Text Classification: Learn from Your Neighbors Ralitsa Angelova , Gerhard Weikum : Max Planck Institute for Informatics Stuhlsatzenhausweg.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Probabilistic Models of Object-Relational Domains
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Link Prediction Topics in Data Mining Fall 2015 Bruno Ribeiro
CPSC 322, Lecture 33Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 33 Nov, 30, 2015 Slide source: from David Page (MIT) (which were.
Lecture 2: Statistical learning primer for biologists
DeepDive Model Dongfang Xu Ph.D student, School of Information, University of Arizona Dec 13, 2015.
Daphne Koller Markov Networks General Gibbs Distribution Probabilistic Graphical Models Representation.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Overview of Relational Markov Networks Colin Evans.
Daphne Koller Overview Conditional Probability Queries Probabilistic Graphical Models Inference.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Markov Networks: Theory and Applications Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
CIS750 – Seminar in Advanced Topics in Computer Science Advanced topics in databases – Multimedia Databases V. Megalooikonomou Link mining ( based on slides.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Learning Deep Generative Models by Ruslan Salakhutdinov
CS 2750: Machine Learning Directed Graphical Models
Qian Liu CSE spring University of Pennsylvania
Boosted Augmented Naive Bayes. Efficient discriminative learning of
Restricted Boltzmann Machines for Classification
Multimodal Learning with Deep Boltzmann Machines
Markov Networks.
Learning with information of features
A Markov Random Field Model for Term Dependencies
Integration and Graphical Models
Lecture 5 Unsupervised Learning in fully Observed Directed and Undirected Graphical Models.
Markov Networks.
Learning Probabilistic Graphical Models Overview Learning Problems.
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
Unifying Variational and GBP Learning Parameters of MNs EM for BNs
Label and Link Prediction in Relational Data
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Markov Networks.
Ping LUO*, Fen LIN^, Yuhong XIONG*, Yong ZHAO*, Zhongzhi SHI^
Presentation transcript:

Discriminative Probabilistic Models for Relational Data Ben Taskar, Pieter Abbeel, Daphne Koller

Tradition statistic classification Methods Dealing with only ‘flat’ data – IID In many supervised learning tasks, entities to be labeled are related to each other in complex way and their labels are not independent This dependence is an important source of information to achieve better classification 4/30/2019 Guohua Hao

Collective Classification Rather than classify each entity separately Simultaneously decide on the class label of all the entities together Explicitly take advantage of the correlation between the labels of related entitiies 4/30/2019 Guohua Hao

Undirected vs. directed graphical models Undirected graphical models do not impose the acyclicity constraint, but directed ones need acyclicity to define a coherent generative model Undirected graphical models are well suited for discriminative training, achieving better classification accuracy over generative training 4/30/2019 Guohua Hao

Our Hypertext Relational Domain Label Label ... ... HasWord1 HasWordk HasWord1 HasWordk Doc Doc From To Link 4/30/2019 Guohua Hao

Schema A set of entity types Attribute of each entity type Content attribute E.X Label attribute E.Y Reference attribute E.R 4/30/2019 Guohua Hao

Provide a set of entities I (E) for each entity type E Instantiation Provide a set of entities I (E) for each entity type E Specify the values of all the attribute of the entities, I.x, I.y, I.r I.r is the instantiation graph, which is call relational skeleton in PRM 4/30/2019 Guohua Hao

Markov Network Qualitative component – Cliques Quantitative component – Potentials 4/30/2019 Guohua Hao

Cliques A set of nodes in the graph G such that for each are connected by an edge in G 4/30/2019 Guohua Hao

Potentials The potential for the clique c defines the compatibility between values of variables in the clique Log-linearly combination of a set of features 4/30/2019 Guohua Hao

Probability in Markov Network Given the values of all nodes in the Markov Network 4/30/2019 Guohua Hao

Conditional Markov Network Specify the probability of a set of target variables Y given a set of conditioning variables X 4/30/2019 Guohua Hao

Relational Markov Network (RMN) Specifies the conditional probability over all the labels of all the entities in the instantiation given the relational structure and the content attributes Extension of the Conditional Markov Networks with a compact definition on a relational data set 4/30/2019 Guohua Hao

Relational clique template F --- a set of entity variables (From) W--- the condition about the attributes of the entity variables (Where) S --- subset of attributes (content and label attribute) of the entity variables (Select) 4/30/2019 Guohua Hao

Relationship to SQL query SELECT doc1.Category,doc2.Category FROM doc1,doc2,Link link WHERE link.From=doc1.key and link.To=doc2.key Doc1 Doc1 Doc2 Link 4/30/2019 Guohua Hao

Potentials Potentials are defined at the level of relational clique template The cliques of the same relational clique template have the same potential functions 4/30/2019 Guohua Hao

Unrolling the RMN Given an instantiation of a relational schema, unroll the RMN as follows Find all the cliques in the unrolled the relational schema where the relational clique templates are applicable The potential of a clique is the same as that of the relational clique template which this clique belongs to 4/30/2019 Guohua Hao

link1 Doc1 Doc2 link2 Doc3 4/30/2019 Guohua Hao

Probability in RMN 4/30/2019 Guohua Hao

4/30/2019 Guohua Hao

Learning RMN Given a set of relational clique templates Estimate feature weight w using conjugate gradient Objective function--Product of likelihood of instantiation and parameter prior Assume a shrinkage prior over feature weights 4/30/2019 Guohua Hao

Learning RMN (Cont’d) The conjugate gradient of the objective function where 4/30/2019 Guohua Hao

Inference in RMN Exact inference Approximate inference Intractable due to the network is very large and densely connected Approximate inference Belief propagation 4/30/2019 Guohua Hao

Experiments WebKB dataset Four CS department websites Five categories (faculty,student,project,course,other) Bag of words on each page Links between pages Experimental setup Trained on three universities Tested on fourth 4/30/2019 Guohua Hao

Flat Models Based only on the text content on the WebPages Incorporate meta-data 4/30/2019 Guohua Hao

Relational model introduce relational clique template over the labels of two pages that are linked Doc1 Doc2 Link 4/30/2019 Guohua Hao

Relational model (Cont’d) relational clique template over the label of section and the label of the pages it is on Relational clique template over the label of the section containing the link and the label of the target page 4/30/2019 Guohua Hao

4/30/2019 Guohua Hao

Discriminative vs. Generative Exit+Naïve Bayes: a complete generative model proposed by Getoor et al Exit+logistic: using logistic regression for the conditional probability distribution of page label given words Link: a fully discriminative training model 4/30/2019 Guohua Hao

Thank You! 4/30/2019 Guohua Hao