Knowledge Graph Embedding

Slides:



Advertisements
Similar presentations
Date: 2014/05/06 Author: Michael Schuhmacher, Simon Paolo Ponzetto Source: WSDM’14 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Knowledge-based Graph Document.
Advertisements

Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
CHAPTER 10: Linear Discrimination
SVM—Support Vector Machines
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Continuous optimization Problems and successes
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
Natural Language Processing Lab, Tsinghua University
Representation learning for Knowledge Bases LivesIn BornIn LocateIn Friendship Nationality Nicole Kidman PerformIn Nationality Sydney Hugh Jackman Australia.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
This week: overview on pattern recognition (related to machine learning)
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Support Vector Machine & Image Classification Applications
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
GAUSSIAN PROCESS FACTORIZATION MACHINES FOR CONTEXT-AWARE RECOMMENDATIONS Trung V. Nguyen, Alexandros Karatzoglou, Linas Baltrunas SIGIR 2014 Presentation:
Group Sparse Coding Samy Bengio, Fernando Pereira, Yoram Singer, Dennis Strelow Google Mountain View, CA (NIPS2009) Presented by Miao Liu July
SVM Support Vector Machines Presented by: Anas Assiri Supervisor Prof. Dr. Mohamed Batouche.
1 Chapter 6. Classification and Prediction Overview Classification algorithms and methods Decision tree induction Bayesian classification Lazy learning.
Non-Isometric Manifold Learning Analysis and an Algorithm Piotr Dollár, Vincent Rabaud, Serge Belongie University of California, San Diego.
Linear Classification with Perceptrons
CSE4334/5334 DATA MINING CSE4334/5334 Data Mining, Fall 2014 Department of Computer Science and Engineering, University of Texas at Arlington Chengkai.
Yue Xu Shu Zhang.  A person has already rated some movies, which movies he/she may be interested, too?  If we have huge data of user and movies, this.
NTU & MSRA Ming-Feng Tsai
Ultra-high dimensional feature selection Yun Li
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
GoRelations: an Intuitive Query System for DBPedia Lushan Han and Tim Finin 15 November 2011
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
A Brief Introduction to Support Vector Machine (SVM) Most slides were from Prof. A. W. Moore, School of Computer Science, Carnegie Mellon University.
1 Bilinear Classifiers for Visual Recognition Computational Vision Lab. University of California Irvine To be presented in NIPS 2009 Hamed Pirsiavash Deva.
Task assignment of interactive Entity resolution 龚赛赛
A Review of Relational Machine Learning for Knowledge Graphs CVML Reading Group Xiao Lin.
Neural Collaborative Filtering
Neural networks and support vector machines
Deep Learning Amin Sobhani.
Deep Compositional Cross-modal Learning to Rank via Local-Global Alignment Xinyang Jiang, Fei Wu, Xi Li, Zhou Zhao, Weiming Lu, Siliang Tang, Yueting.
Learning Mid-Level Features For Recognition
Wenhan Xiong, Thien Hoang, William Wang Department of Computer Science
Saisai Gong, Wei Hu, Yuzhong Qu
Attention Is All You Need
Source: Procedia Computer Science(2015)70:
Machine Learning Basics
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Torch 02/27/2018 Hyeri Kim Good afternoon, everyone. I’m Hyeri. Today, I’m gonna talk about Torch.
Support Vector Machines Introduction to Data Mining, 2nd Edition by
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Pattern Recognition CS479/679 Pattern Recognition Dr. George Bebis
Variational Knowledge Graph Reasoning
Learning with information of features
Knowledge Base Completion
Goodfellow: Chapter 14 Autoencoders
COSC 4335: Other Classification Techniques
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
Computer Vision Chapter 4
Graph and Tensor Mining for fun and profit
Paradigms for Computational Nucleic Acid Design
Improving Knowledge Graph Embedding Using Simple Constraints
Leverage Consensus Partition for Domain-Specific Entity Coreference
Machine learning overview
COSC 4368 Machine Learning Organization
Neural networks (3) Regularization Autoencoder
Autoencoders Supervised learning uses explicit labels/correct output in order to train a network. E.g., classification of images. Unsupervised learning.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
NAACL‘18 Dai Quoc Nguyen, Tu Dinh Nguyen, Dat Quoc Nguyen, Dinh Phung
Linear Discrimination
Graph Attention Networks
Adversarial Personalized Ranking for Recommendation
Peng Cui Tsinghua University
Goodfellow: Chapter 14 Autoencoders
Presentation transcript:

Knowledge Graph Embedding 龚赛赛 saisaigong@gmail.com 2015-01-26

Reference Zhen Wang, Jianwen Zhang, Jianlin Feng, Zheng Chen. Knowledge Graph Embedding by Translating on Hyperplanes. AAAI14 Antoine Bordes, Nicolas Usunier, Alberto Garcia-Dur´an. Translating Embeddings for Modeling Multi-relational Data. NIPS13 Antoine Bordes, JasonWeston, Ronan Collobert, Yoshua Bengio. Learning Structured Embeddings of Knowledge Bases. AAAI11

Contents Introduction Overview Algorithm framework Applications

Introduction Great achievements in building large scale Knowledge graphs Widely usage of several KGs such as WordNet, Freebase, Dbpedia NLP, collaborative filtering, computer vision problem, etc However, the general paradigm to support computing is still not clear Two major difficulties KG: symbolic and logical vs Application: numerical computing in continuous spaces difficult to aggregate global knowledge over a graph E.g. reasoning neither tractable nor robust

Introduction Embed a knowledge graph into a continuous vector space while preserving certain properties of the original graph Also apply to multi-relational data

Overview KG: a set of triplets (h,r,t ) h: head (subject) r: relation (predicate) t: tail (object) Each entity is represented by a k-dimensional vector h (t) A scoring function f r(h,t): measure the plausibility of a triplet (h,r,t) in the embedding space Imply a transformation r on the pair of the entities which characterizes the relation r

Overview Scoring functions Simple difference Translation Affine General linear Bilinear Nonlinear transformations

Overview Scoring functions Simple difference Translation Affine General linear Bilinear Nonlinear transformations V(马德里)-v(西班牙)+v(法国)=v(巴黎)

Overview Translation Linear Translation

Overview TransE (nips13) TransH(aaai14)

Overview Distant(aaai11) More complex General linear Bilinear (tensor) NTN (neural network) …

Algorithm framwork Optimization object: minimize the scores of the true triplets and maximize the scores of the negative triplets Training Algorithm: stochastic gradient descent (minibatch) Negative triplets (∆’(h,r,t)): corrupt true triplets, e.g. (h,r,t)->(h,r,t’) or (h’,r,t) Reduce false negative triplets TransH

Algorithm framework 1. 初始化vectors 2. 每次随机挑选出一组true triplets

Applications Link prediction Triplets classification Complete a triple (h,r,t) with h or t missing e.g. predict the RDF classes of an entity (rdf:type) Triplets classification Confirm whether (h,r,t) is correct or not Relational fact extraction from text Learn on riplets from text-> induce new facts Entity Relatedness e.g. Learning taxonomies

Performance Dataset Link prediction Rank candidate objects/subjects Evaluation metric: averaged rank, Hits@10

Performance

Performance sparse dense

Thank you for you attention !