Constructing Knowledge Graph from Unstructured Text Image Source: www.ibm.com/smarterplanet/us/en/ibmwatson/ Kundan Kumar Siddhant Manocha.

Slides:



Advertisements
Similar presentations
Machine learning continued Image source:
Advertisements

Word sense disambiguation and information retrieval Chapter 17 Jurafsky, D. & Martin J. H. SPEECH and LANGUAGE PROCESSING Jarmo Ritola -
Deep Learning in NLP Word representation and how to use it for Parsing
Deep Learning.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
Word/Doc2Vec for Sentiment Analysis
Sentiment Lexicon Creation from Lexical Resources BIS 2011 Bas Heerschop Erasmus School of Economics Erasmus University Rotterdam
An Overview of Text Mining Rebecca Hwa 4/25/2002 References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37 th Annual Meeting of.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Distributed Representations of Sentences and Documents
Linguistic Regularities in Sparse and Explicit Word Representations
Yang-de Chen Tutorial: word2vec Yang-de Chen
Information Retrieval in Practice
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
L’età della parola Giuseppe Attardi Dipartimento di Informatica Università di Pisa ESA SoBigDataPisa, 24 febbraio 2015.
Open Information Extraction using Wikipedia
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
1 Learning Sub-structures of Document Semantic Graphs for Document Summarization 1 Jure Leskovec, 1 Marko Grobelnik, 2 Natasa Milic-Frayling 1 Jozef Stefan.
Efficient Estimation of Word Representations in Vector Space
Indirect Supervision Protocols for Learning in Natural Language Processing II. Learning by Inventing Binary Labels This work is supported by DARPA funding.
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio Supervised Relation Extraction.
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Omer Levy Yoav Goldberg Ido Dagan Bar-Ilan University Israel
Cold Start Problem in Movie Recommendation JIANG CAIGAO, WANG WEIYAN Group 20.
Deep Visual Analogy-Making
4. Relationship Extraction Part 4 of Information Extraction Sunita Sarawagi 9/7/2012CS 652, Peter Lindes1.
Learning Kernel Classifiers 1. Introduction Summarized by In-Hee Lee.
Ganesh J, Soumyajit Ganguly, Manish Gupta, Vasudeva Varma, Vikram Pudi
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning Ronan Collobert Jason Weston Presented by Jie Peng.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
RELATION EXTRACTION, SYMBOLIC SEMANTICS, DISTRIBUTIONAL SEMANTICS Heng Ji Oct13, 2015 Acknowledgement: distributional semantics slides from.
Multi-Class Sentiment Analysis with Clustering and Score Representation Yan Zhu.
A Tutorial on ML Basics and Embedding Chong Ruan

Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Efficient Estimation of Word Representations in Vector Space By Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean. Google Inc., Mountain View, CA. Published.
Short Text Similarity with Word Embedding Date: 2016/03/28 Author: Tom Kenter, Maarten de Rijke Source: CIKM’15 Advisor: Jia-Ling Koh Speaker: Chih-Hsuan.
Word Sense Disambiguation Algorithms in Hindi
An Integrated Approach for Relation Extraction from Wikipedia Texts Yulan Yan Yutaka Matsuo Mitsuru Ishizuka The University of Tokyo WWW 2009.
DeepWalk: Online Learning of Social Representations
Medical Semantic Similarity with a Neural Language Model Dongfang Xu School of Information Using Skip-gram Model for word embedding.
Distributed Representations for Natural Language Processing
Cross-lingual Dataless Classification for Many Languages
Sentiment analysis using deep learning methods
Semi-Supervised Clustering
Korean version of GloVe Applying GloVe & word2vec model to Korean corpus speaker : 양희정 date :
Deep Learning for Bacteria Event Identification
A Brief Introduction to Distant Supervision
Cross-lingual Dataless Classification for Many Languages
Intro to NLP and Deep Learning
Distributed Representations of Words and Phrases and their Compositionality Presenter: Haotian Xu.
Deep learning and applications to Natural language processing
Efficient Estimation of Word Representation in Vector Space
Word2Vec CS246 Junghoo “John” Cho.
Distributed Representation of Words, Sentences and Paragraphs
Jun Xu Harbin Institute of Technology China
Lei Sha, Jing Liu, Chin-Yew Lin, Sujian Li, Baobao Chang, Zhifang Sui
Deep Learning Hierarchical Representations for Image Steganalysis
Word Embedding Word2Vec.
Word embeddings based mapping
Enriching Taxonomies With Functional Domain Knowledge
Word2Vec.
Word embeddings (continued)
CSE 291G : Deep Learning for Sequences
Motivation It can effectively mine multi-modal knowledge with structured textural and visual relationships from web automatically. We propose BC-DNN method.
Presentation transcript:

Constructing Knowledge Graph from Unstructured Text Image Source: Kundan Kumar Siddhant Manocha

MOTIVATION Image Source: KDD 2014 Tutorial on Constructing and Mining Web-scale Knowledge Graphs, New York

MOTIVATION

Image Source: KDD 2014 Tutorial on Constructing and Mining Web-scale Knowledge Graphs, New York MOTIVATION

PROBLEM STATEMENT

KNOWLEDGE GRAPH

KNOWLEDGE GRAPH

QUESTION ANSWERING

EXISTING KNOWLEDGE BASES Image Source: KDD 2014 Tutorial on Constructing and Mining Web-scale Knowledge Graphs, New York

EXISTING KNOWLEDGE BASES Supervised Models: ◦Learn classifiers from +/- examples, typical features: context words + POS, dependency path between entities, named entity tags ◦Require large number of tagged training examples ◦Cannot be generalized Semi-Supervised Models: ◦Bootstrap Algorithms: Use seed examples to learn initial set of relations ◦Generate +ve/-ve examples to learn a classifier ◦Learn more relations using this classifier Distant Supervision: ◦Existing knowledge base + unlabeled text generate examples ◦Learn models using this set of relations

OUR APPROACH Bootstrapping Relations using Distributed Word Vector Embedding 1) Word that occur in similar context lie close together in the word embedding space. 2) Word Vectors is semantically consistent and capture many linguistic properties (like 'capital city', 'native language', 'plural relations') 3) Obtain word vectors from unstructured text ( using Google word2vec, Glove, etc ) 4) Exploit the properties of the manifold to obtain binary relations between entities

ALGORITHM Image Source: KDD 2014 Tutorial on Constructing and Mining Web-scale Knowledge Graphs, New York

SIMILARITY METRIC Image Source:A survey on relation extraction, Nguyen Bach, Carnegie Mellon University

KERNEL BASED APPROACHES

DEPENDENCY KERNELS 1.Actual Sentences 2. Dependency Graph Kernel: K(x,y)=3×1×1×1×2×1×3 = 18 3.Kernel Computation Image Source:A Shortest Path Dependency Kernel for Relation Extraction,Mooney,et al

PRELIMINARY RESULTS Word Vector Embedding: Wikipedia Corpus

PRELIMINARY RESULTS (wikipedia corpus) Seed Examples for capital relationship Positive relations learnt Negative Relations learnt

PRELIMINARY RESULTS (google news corpus) Seed Examples Positive Relations Learned Negative Relations Learned

References 1)Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. Linguistic Regularities in Continuous Space Word Representations. In Proceedings of NAACL HLT, ) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeffrey Dean. Distributed Representations of Words and Phrases and their Compositionality. In Proceedings of NIPS, 2013.Distributed Representations of Words and Phrases and their Compositionality 3) Eugene Agichtein Luis Gravano. Snowball: Extracting Relations from Large Plain-Text Collections. In Proceedings of the fifth ACM conference on Digital libraries, June 2000

Questions!

CBOW MODEL input vector represented as 1-of-V encoding Linear sum of input vectors are projected onto the projection layer Hierarchical Softmax layer is used to ensure that the weights in the output layer are between 0<=p<=1 Weights learnt using back- propagation The projection matrix from the projection layer to the hidden layer give the word vector embeddings Image Source: Linguistic Regularities in Continuous Space Word Representations,Mikolov,et.al 2013

WORD VECTOR MODEL

KERNEL BASED APPROACHES Image Source:Kernel Methods for Relation Extraction,Zelenko,et al