Knowledge Representation and Reasoning

Slides:



Advertisements
Similar presentations
Multi-label Relational Neighbor Classification using Social Context Features Xi Wang and Gita Sukthankar Department of EECS University of Central Florida.
Advertisements

GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
Research topics Semantic Web - Spring 2007 Computer Engineering Department Sharif University of Technology.
Building Knowledge-Driven DSS and Mining Data
1 A Topic Modeling Approach and its Integration into the Random Walk Framework for Academic Search 1 Jie Tang, 2 Ruoming Jin, and 1 Jing Zhang 1 Knowledge.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Face Detection using the Viola-Jones Method
1 Zi Yang, Wei Li, Jie Tang, and Juanzi Li Knowledge Engineering Group Department of Computer Science and Technology Tsinghua University, China {yangzi,
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
1 A Discriminative Approach to Topic- Based Citation Recommendation Jie Tang and Jing Zhang Presented by Pei Li Knowledge Engineering Group, Dept. of Computer.
X-Informatics Web Search; Text Mining B 2013 Geoffrey Fox Associate Dean for.
Some working definitions…. ‘Data Mining’ and ‘Knowledge Discovery in Databases’ (KDD) are used interchangeably Data mining = –the discovery of interesting,
1 Yang Yang *, Yizhou Sun +, Jie Tang *, Bo Ma #, and Juanzi Li * Entity Matching across Heterogeneous Sources *Tsinghua University + Northeastern University.
On Node Classification in Dynamic Content-based Networks.
DATA MINING WITH CLUSTERING AND CLASSIFICATION Spring 2007, SJSU Benjamin Lam.
M. Wang, T. Xiao, J. Li, J. Zhang, C. Hong, & Z. Zhang (2014)
1 Yang Yang *, Yizhou Sun +, Jie Tang *, Bo Ma #, and Juanzi Li * Entity Matching across Heterogeneous Sources *Tsinghua University + Northeastern University.
TWC Illuminate Knowledge Elements in Geoscience Literature Xiaogang (Marshall) Ma, Jin Guang Zheng, Han Wang, Peter Fox Tetherless World Constellation.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Mining Advisor-Advisee Relationships from Research Publication.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
DeepWalk: Online Learning of Social Representations
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Multi-Modal Bayesian Embeddings for Learning Social Knowledge Graphs Zhilin Yang 12, Jie Tang 1, William W. Cohen 2 1 Tsinghua University 2 Carnegie Mellon.
Big data classification using neural network
MINING DEEP KNOWLEDGE FROM SCIENTIFIC NETWORKS
Deep Compositional Cross-modal Learning to Rank via Local-Global Alignment Xinyang Jiang, Fei Wu, Xi Li, Zhou Zhao, Weiming Lu, Siliang Tang, Yueting.
Efficient Image Classification on Vertically Decomposed Data
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Cross-lingual Knowledge Linking Across Wiki Knowledge Bases
Collective Network Linkage across Heterogeneous Social Platforms
Aspect-based sentiment analysis
Generating Natural Answers by Incorporating Copying and Retrieving Mechanisms in Sequence-to-Sequence Learning Shizhu He, Cao liu, Kang Liu and Jun Zhao.
Fenglong Ma1, Jing Gao1, Qiuling Suo1
Deep Learning based Machine Translation
Efficient Image Classification on Vertically Decomposed Data
Community Distribution Outliers in Heterogeneous Information Networks
Presentation 王睿.
Weakly Learning to Match Experts in Online Community
Towards Understanding the Invertibility of Convolutional Neural Networks Anna C. Gilbert1, Yi Zhang1, Kibok Lee1, Yuting Zhang1, Honglak Lee1,2 1University.
Knowledge Base Completion
ece 627 intelligent web: ontology and beyond
Disambiguation Algorithm for People Search on the Web
Table Cell Search for Question Answering Huan Sun
iSRD Spam Review Detection with Imbalanced Data Distributions
MEgo2Vec: Embedding Matched Ego Networks for User Alignment Across Social Networks Jing Zhang+, Bo Chen+, Xianming Wang+, Fengmei Jin+, Hong Chen+, Cuiping.
Introduction to Machine Reading Comprehension
Deep Cross-media Knowledge Transfer
Resource Recommendation for AAN
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Example: Academic Search
Asymmetric Transitivity Preserving Graph Embedding
Graph-based Security and Privacy Analytics via Collective Classification with Joint Weight Learning and Propagation Binghui Wang, Jinyuan Jia, and Neil.
View Inter-Prediction GAN: Unsupervised Representation Learning for 3D Shapes by Learning Global Shape Memories to Support Local View Predictions 1,2 1.
GANG: Detecting Fraudulent Users in OSNs
Graph Neural Networks Amog Kamsetty January 30, 2019.
边缘检测年度进展概述 Ming-Ming Cheng Media Computing Lab, Nankai University
Actively Learning Ontology Matching via User Interaction
Rachit Saluja 03/20/2019 Relation Extraction with Matrix Factorization and Universal Schemas Sebastian Riedel, Limin Yao, Andrew.
Keshav Balasubramanian
Multiple DAGs Learning with Non-negative Matrix Factorization
Building Dynamic Knowledge Graphs From Text Using Machine Reading Comprehension Rajarshi Das, Tsendsuren Munkhdalai, Xingdi Yuan, Adam Trischler, Andrew.
Topic: Semantic Text Mining
Modeling IDS using hybrid intelligent systems
Bug Localization with Combination of Deep Learning and Information Retrieval A. N. Lam et al. International Conference on Program Comprehension 2017.
Representation Learning on Networks
Peng Cui Tsinghua University
--WWW 2010, Hongji Bao, Edward Y. Chang
Network Embedding under Partial Monitoring for Evolving Networks
Jifan Yu, Chenyu Wang, Gan Luo,
Presentation transcript:

Knowledge Representation and Reasoning Jie Tang Department of Computer Science and Technology Tsinghua University The slides can be downloaded at http://keg.cs.tsinghua.edu.cn/jietang

Knowledge Graph “Knowledge graph” was used by Google 2012 Knowledge engineering, expert system CYC: the world's longest-lived AI project (1985)

Knowledge & AI Big Data Knowledge Intelligence * Two keys for AI: Knowledge Base + Intelligent Algorithm Tim Berners Lee Father of WWW Turing Award Edward Feigenbaum Father of KB Turing Award How to populate a semantic-based profile database for researchers? Big Data Knowledge Intelligence [1] J. Tang, J. Zhang, L. Yao, J. Li, L. Zhang, and Z. Su. ArnetMiner: Extraction and Mining of Academic Social Networks. KDD’08. pp.990-998.

Example: Multi-hop QA Quality Café Los Angeles Old School Alessandro Moschitti The Quality Cafe is a now-defunct diner in Los Angeles, California. The restaurant has appeared as a location featured in a number of Hollywood films, including Old School, Gone in 60 Seconds, ... Los Angeles is the most populous city in California, the second most populous city in the United States, after New York City, and the third most populous city in North America. Alessandro Moschitti is a professor of the CS Department of the University of Trento, Italy. He is currently a Principal Research Scientist of the Qatar Computing Research Institute (QCRI) Old School Todd Phillips Tsinghua University Todd Phillips is an American director, producer, screenwriter, and actor. He is best known for writing and directing films, including Road Trip (2000), Old School (2003), Starsky & Hutch (2004), and The Hangover Trilogy. Tsinghua University is a major research university in Beijing and dedicated to academic excellence and global development. Tsinghua is perennially ranked as one of the top academic institutions in China, Asia, and worldwide... Old School is a 2003 American comedy film released by Dream Works Pictures and The Montecito Picture Company and directed by Todd Phillips.

Related work: DrQA framework DrQA [1] proposed a popular two-stage framework for open-domain QA. Document Retriever: Retrieve 5 related Wikipedia articles given any question. Document Reader: Predict the span of answer with (usually deep learning) models. Chen, Danqi, et al. "Reading wikipedia to answer open-domain questions." arXiv preprint arXiv:1704.00051 (2017).

Challenge 1: Myopic Retrieval Problem Document A Old School is a 2003 American comedy film released by DreamWorks Pictures and The Montecito Picture Company… Document B Many directors shoot scenes in Hollywood, Los Angeles, which is notable as the home of the U.S. film industry. document B share more common words with the question.

Related work: Reading Comprehension models The Document Readers in DrQA framework are targeted at single-paragraph reading comprehension QA, whose benchmark is SQuAD[1] dataset. BiDAF BERT XLNet Rajpurkar, Pranav, et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text. EMNLP’’16.

Challenge 2: Explainability Most RC models can only be seen as black-boxes: For verification, the user needs: reasoning path (graph) supporting facts (sentences) at each hop Other possible answers and their reasoning paths for comparison

Ideal Results Imagine a scenario where you have a wikipedia to look up for any entity, and how can you answer such a question?(click) Question… We will first search quality café and los angeles which seems useful for answering the question. Okay there are two quality café but we find CLUES in the middle paragraph. There are two films featured here. And then we read the paragraph in wiki about the two films and find out there directors. Now, we construct a graph step by step, which includes all possible answers and userful information. And our brain will analyze the whole information and chooses out the final answer. We name this graph “cognitive graph”. // We get the final answer, and review the whole process, two non-trivial things were done by our brain. First, we identified useful entities from text such as “Old School” “Gone in 60 Seconds” “Todd Phillips”…This process is intuitively and simple. Then, we should reason out the final answer Todd Philips from other distractors for example Dominic Sena. This problem is much harder.

Cognitive Psychology Framework System 1 Intuitive System 2 Analytic Dual Process Theory Jonathan St BT Evans. 1984. Heuristic and analytic processes in reasoning. British Journal of Psychology, 75(4):451–468.

How to do the reasoning?

Cognitive Graph System 1: System 2: Knowledge expansion by association in text when reading System 2: Decision making w/ all the information Dual process theory in cognition science says: System 1 Intuitive System 2 Analytic

CogQA: Cognitive Graph for QA An iterative framework corresponding to dual process theory System 1 extract entities to build the cognitive graph generate semantic vectors for each node System 2 Do reasoning based on semantic vectors and graph Feed clues to System 1 to extract next-hop entities System 1 extract Question Cognitive Graph input Quality café Los Angeles …location featured in a number of Hollywood films, including Old School, Gone in 60 Seconds… clues Gone in 60 seconds Old school input System 2 Dominic Sena Todd Phillips predict

Cognitive Graph: DL + Dual Process Theory explicit decision implicit knowledge expansion M. Ding, C. Zhou, Q. Chen, H. Yang, and J. Tang. Cognitive Graph for Multi-Hop Reading Comprehension at Scale. ACL’19.

System 1: BERT Implementation Extract top-k next-hop entities and answer candidates respectively Predict the start and end probabilities of each position Generate semantic vectors for entities based on their documents Take the 0-th probability as negative threshold Ignore the spans whose start probabilities are small than the negative threshold

System 2: the GNN implementation At each step, hidden representations X for nodes are updated according to the propagation rules: Predictor F is a two-layer MLP, which predicts the final answer based on hidden representations X:

CogQA Algorithm and Training Preprocess Mark the spans of next-hop entities and answer by fuzzy matching Golden cognitive graph and negative (hop and span) nodes Train Task #1 (System 1): Next-hop spans and answer spans extraction Train Task #2 (System 1 & 2): Graph convolution and answer prediction

Performance HotpotQA is a dataset with leaderboard similar to SQuAD CogQA ranked 1st from 21, Feb to 15, May (nearly 3 month) Two things to note: Yang-IR: improved retrieval. We improve the retrieval process of Yang et al. for better comparison. CogQA-onlyQ is designed that it only starts with the entities that appear in the question. Free of elaborate retrieval methods, this setting can be regarded as a natural thinking pattern of human being, in which only explicit and reliable relations are needed in reasoning. CogQA-onlyQ still outperforms all the baselines ** Code available at https://github.com/THUDM/CogQA

Reasoning Power CogQA Performs much better on question with more hops!

Case Study Tree-shape Cognitive Graph Users can verify the answer by comparing it with another possible reasoning chain. “Upper House” in the question is similar to “Senate” not “House of Representative” We show how the cognitive graph clearly explains complex reasoning processes in our experiments. The cognitive graph highlights the heart of the question in case (1) – i.e., to choose between the number of members in two houses. CogQA makes the right choice based on semantic similarity between “Senate” and “upper house”. Case (2) illustrates that the robustness of answering can be boosted by exploring parallel reasoning paths. In Case (3) our model finally gets the answer “Marijus Adomaitis” while the annotated ground truth is “Ten Walls”. However, when backtracking the reasoning process in cognitive graph, we find that the model has already reached “Ten Walls” and answers with his real name, which is acceptable and even more accurate. This is why we need explainability.

Case Study DAG-shape Cognitive Graph Multiple supporting facts provides richer information, increasing the credibility of the answer.

Case Study In this case, CogQA gives the answer “Marijus Adomaitis” while the annotated ground truth is “Ten Walls”. By examining the cognitive graph, we find that Ten Walls is just the stage name of Marijus Adomaitis! Without cognitive graphs, black-box models cannot achieve it.

CogDL —A Toolkit for ​Deep Learning on Graphs ** Code available at https://keg.cs.tsinghua.edu.cn/cogdl/

CogDL —A Toolkit for ​Deep Learning on Graphs CogDL is a graph representation learning toolkit that allows researchers and developers to train baseline or custom models for node classification, link prediction and other tasks on graphs. It provides implementations of several models, including: non-GNN Baselines like Deepwalk, LINE, NetMF, GNN Baselines like GCN, GAT, Graphsage.

CogDL features CogDL supports these features: Sparsification: fast network embedding on large-scale networks with tens of millions of nodes Arbitrary: dealing with different graph structures: attributed, multiplex and heterogeneous networks Distributed: multi-GPU training on one machine or across multiple machines Extensible: easily register new datasets, models, criterions and tasks Homepage: https://keg.cs.tsinghua.edu.cn/cogdl/ CogDL support these features: Sparsification, Arbitrary, Distributed, Extensible.

CogDL Leaderboards Multi-label Node Classification The following leaderboard is built from unsupervised learning with multi-label node classification setting. CogDL provides several downstream tasks including node classification (with or without node attributes), link prediction (with or without attributes, heterogeneous or not). These leaderboards maintain state-of-the-art results and benchmarks on these tasks. The following leaderboard is built from unsupervised learning with multi-label node classification setting. We run all algorithms on several real-world datasets and report the sorted experimental results. ProNE performs best in all algorithms. http://keg.cs.tsinghua.edu.cn/cogdl/node-classification.html

CogDL Leaderboards Node Classification with Attributes. We also built a leaderboards on node classification with attributes. We compare several popular graph neural network methods like GCN, GAT on several popular datasets like Cora, Citeseer and Pubmed. http://keg.cs.tsinghua.edu.cn/cogdl/node-classification.html

Leaderboards: Link Prediction For link prediction, we adopt Area Under the Receiver Operating Characteristic Curve (ROC AUC), standard evaluation metric AUC and F1- score, which represents the probability that vertices in a random unobserved link are more similar than those in a random nonexistent link. We evaluate these measures while removing 15 percents of edges on these dataset. We repeat our experiments for 10 times and report the three matrices in order. http://keg.cs.tsinghua.edu.cn/cogdl/link-prediction.html

Join us Feel free to join us with the three following ways: add your data into the leaderboard add your algorithm into the leaderboard add your algorithm into the toolkit Note that, in order to collaborate with us, you may refer to the documentation in our homepage. You are welcome to join us.

Cognitive Graph: DL + Dual Process Theory explicit decision implicit knowledge expansion M. Ding, C. Zhou, Q. Chen, H. Yang, and J. Tang. Cognitive Graph for Multi-Hop Reading Comprehension at Scale. ACL’19.

Related Publications Ming Ding, Chang Zhou, Qibin Chen, Hongxia Yang, and Jie Tang. Cognitive Graph for Multi-Hop Reading Comprehension at Scale. ACL’19. Jie Zhang, Yuxiao Dong, Yan Wang, Jie Tang, and Ming Ding. ProNE: Fast and Scalable Network Representation Learning. IJCAI’19. Yukuo Cen, Xu Zou, Jianwei Zhang, Hongxia Yang, Jingren Zhou and Jie Tang. Representation Learning for Attributed Multiplex Heterogeneous Network. KDD’19. Fanjin Zhang, Xiao Liu, Jie Tang, Yuxiao Dong, Peiran Yao, Jie Zhang, Xiaotao Gu, Yan Wang, Bin Shao, Rui Li, and Kuansan Wang. OAG: Toward Linking Large-scale Heterogeneous Entity Graphs. KDD’19. Yifeng Zhao, Xiangwei Wang, Hongxia Yang, Le Song, and Jie Tang. Large Scale Evolving Graphs with Burst Detection. IJCAI’19. Yu Han, Jie Tang, and Qian Chen. Network Embedding under Partial Monitoring for Evolving Networks. IJCAI’19. Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Chi Wang, Kuansan Wang, and Jie Tang. NetSMF: Large-Scale Network Embedding as Sparse Matrix Factorization. WWW'19. Jiezhong Qiu, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, and Jie Tang. DeepInf: Modeling Influence Locality in Large Social Networks. KDD’18. Jiezhong Qiu, Yuxiao Dong, Hao Ma, Jian Li, Kuansan Wang, and Jie Tang. Network Embedding as Matrix Factorization: Unifying DeepWalk, LINE, PTE, and node2vec. WSDM’18. Jie Tang, Jing Zhang, Limin Yao, Juanzi Li, Li Zhang, and Zhong Su. ArnetMiner: Extraction and Mining of Academic Social Networks. KDD’08. For more, please check here http://keg.cs.tsinghua.edu.cn/jietang

Thank you! Collaborators: Ming Ding, Qibin Chen, et al. (THU) Hongxiao Yang, Chang Zhou (Alibaba) Jie Tang, KEG, Tsinghua U http://keg.cs.tsinghua.edu.cn/jietang Download all data & Codes https://keg.cs.tsinghua.edu.cn/cogdl/