Modeling Scientific Impact with Topical Influence Regression James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine.

Slides:



Advertisements
Similar presentations
Topic models Source: Topic models, David Blei, MLSS 09.
Advertisements

Information retrieval – LSI, pLSI and LDA
Statistical Machine Translation Part II: Word Alignments and EM Alexander Fraser ICL, U. Heidelberg CIS, LMU München Statistical Machine Translation.
Hierarchical Dirichlet Processes
Simultaneous Image Classification and Annotation Chong Wang, David Blei, Li Fei-Fei Computer Science Department Princeton University Published in CVPR.
Title: The Author-Topic Model for Authors and Documents
Text-Based Measures of Document Diversity Date : 2014/02/12 Source : KDD’13 Authors : Kevin Bache, David Newman, and Padhraic Smyth Advisor : Dr. Jia-Ling,
An Introduction to LDA Tools Kuan-Yu Chen Institute of Information Science, Academia Sinica.
Statistical Topic Modeling part 1
MICHAEL PAUL AND ROXANA GIRJU UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN A Two-Dimensional Topic-Aspect Model for Discovering Multi-Faceted Topics.
Decoupling Sparsity and Smoothness in the Discrete Hierarchical Dirichlet Process Chong Wang and David M. Blei NIPS 2009 Discussion led by Chunping Wang.
Unsupervised and Weakly-Supervised Probabilistic Modeling of Text Ivan Titov April TexPoint fonts used in EMF. Read the TexPoint manual before.
Generative Topic Models for Community Analysis
Statistical Models for Networks and Text Jimmy Foulds UCI Computer Science PhD Student Advisor: Padhraic Smyth.
Topic Modeling with Network Regularization Md Mustafizur Rahman.
Stochastic Collapsed Variational Bayesian Inference for Latent Dirichlet Allocation James Foulds 1, Levi Boyles 1, Christopher DuBois 2 Padhraic Smyth.
Latent Dirichlet Allocation a generative model for text
Language Models for TR Rong Jin Department of Computer Science and Engineering Michigan State University.
British Museum Library, London Picture Courtesy: flickr.
Multiscale Topic Tomography Ramesh Nallapati, William Cohen, Susan Ditmore, John Lafferty & Kin Ung (Johnson and Johnson Group)
Scalable Text Mining with Sparse Generative Models
Topic models for corpora and for graphs. Motivation Social graphs seem to have –some aspects of randomness small diameter, giant connected components,..
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Semantic History Embedding in Online Generative Topic Models Pu Wang (presenter) Authors: Loulwah AlSumait Daniel Barbará
Introduction to Machine Learning for Information Retrieval Xiaolong Wang.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
Modeling Documents by Combining Semantic Concepts with Unsupervised Statistical Learning Author: Chaitanya Chemudugunta America Holloway Padhraic Smyth.
Online Learning for Latent Dirichlet Allocation
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior National Business Center contract number.
Topic Modelling: Beyond Bag of Words By Hanna M. Wallach ICML 2006 Presented by Eric Wang, April 25 th 2008.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
NUDT Machine Translation System for IWSLT2007 Presenter: Boxing Chen Authors: Wen-Han Chao & Zhou-Jun Li National University of Defense Technology, China.
Style & Topic Language Model Adaptation Using HMM-LDA Bo-June (Paul) Hsu, James Glass.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Integrating Topics and Syntax -Thomas L
Summary We propose a framework for jointly modeling networks and text associated with them, such as networks or user review websites. The proposed.
A Model for Learning the Semantics of Pictures V. Lavrenko, R. Manmatha, J. Jeon Center for Intelligent Information Retrieval Computer Science Department,
Latent Dirichlet Allocation D. Blei, A. Ng, and M. Jordan. Journal of Machine Learning Research, 3: , January Jonathan Huang
Probabilistic Models for Discovering E-Communities Ding Zhou, Eren Manavoglu, Jia Li, C. Lee Giles, Hongyuan Zha The Pennsylvania State University WWW.
Xinran He1, Theodoros Rekatsinas2,
Storylines from Streaming Text The Infinite Topic Cluster Model Amr Ahmed, Jake Eisenstein, Qirong Ho Alex Smola, Choon Hui Teo, Eric Xing Carnegie Mellon.
Topic Models Presented by Iulian Pruteanu Friday, July 28 th, 2006.
Topic Modeling using Latent Dirichlet Allocation
Latent Dirichlet Allocation
1 A Biterm Topic Model for Short Texts Xiaohui Yan, Jiafeng Guo, Yanyan Lan, Xueqi Cheng Institute of Computing Technology, Chinese Academy of Sciences.
School of Computer Science 1 Information Extraction with HMM Structures Learned by Stochastic Optimization Dayne Freitag and Andrew McCallum Presented.
1 Toward Metrics of Design Automation Research Impact Andrew B. Kahng ‡†, Mulong Luo †, Gi-Joon Nam 1, Siddhartha Nath †, David Z. Pan 2 and Gabriel Robins.
1 Minimum Error Rate Training in Statistical Machine Translation Franz Josef Och Information Sciences Institute University of Southern California ACL 2003.
Towards Total Scene Understanding: Classification, Annotation and Segmentation in an Automatic Framework N 工科所 錢雅馨 2011/01/16 Li-Jia Li, Richard.
Link Distribution on Wikipedia [0407]KwangHee Park.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
How can we maintain an error bound? Settle for a “per-step” bound What’s the probability of a mistake at each step? Not cumulative, but Equal footing with.
Concept-Based Analysis of Scientific Literature Chen-Tse Tsai, Gourab Kundu, Dan Roth UIUC.
Review: Review: Translating without in-domain corpus: Machine translation post-editing with online learning techniques Antonio L. Lagarda, Daniel Ortiz-Martínez,
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
Topic Modeling for Short Texts with Auxiliary Word Embeddings
Sentiment analysis algorithms and applications: A survey
Online Multiscale Dynamic Topic Models
J. Zhu, A. Ahmed and E.P. Xing Carnegie Mellon University ICML 2009
Bayesian Inference for Mixture Language Models
John Lafferty, Chengxiang Zhai School of Computer Science
Topic models for corpora and for graphs
Resource Recommendation for AAN
Michal Rosen-Zvi University of California, Irvine
Dynamic Supervised Community-Topic Model
Topic Models in Text Processing
Hierarchical Relational Models for Document Networks
Presentation transcript:

Modeling Scientific Impact with Topical Influence Regression James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine

Exploring a New Scientific Area 2

3 Which are the most important articles?

Exploring a New Scientific Area 4 What are the influence relationships between articles?

Outline Background: Modeling scientific impact, topic models Metric: Topical Influence Model: Topical Influence Regression Inference Algorithm Experimental Results 5

Can’t We Just Use Citation Counts? Many citations are made out of “politeness, policy or piety” [Ziman, 1968]. Mentioned (A) in passing Built upon the ideas of (B) Which article is more influential? Article (A)Article (B) 6

Enter: Natural Language Processing Use NLP techniques to exploit textual information in conjunction with citation information Using this extra information, we should be able to gain a deeper understanding of scientific impact than simple citation counts 7

Previous Approaches Traditional Bibliometrics – Citation counts, journal impact factors, H-Index Graph-based – PageRank on the citation graph – PageRank on an article similarity graph (Lin, 2008) Supervised Machine Learning – Classifying citation function (Teufel et al., 2006) NLP / Topic Models – Dietz et al. (2007), Gerrish & Blei (2010), Nallapati et al. (2011) … 8

Our Approach A metric arising from a generative probabilistic model for scientific corpora Fully unsupervised Exploits both textual content and the citation graph Recovers both node-level and edge-level influence scores A flexible, extensible regression framework 9

Latent Dirichlet Allocation Topic Models Topic models are a bag of words approach to modeling text corpora Topics are distributions over words Every document has a distribution over topics, with a Dirichlet prior Every word is assigned a latent topic, which it is assumed to be drawn from. 10

Latent Dirichlet Allocation and Polya Urns For each document – Place colored balls in that document’s urn, where each color is associated with a topic, and α is the Dirichlet prior on the distribution over topics. – For each word Draw a ball from the urn, observe its color k Draw the word token from topic k Place the ball back, along with a new ball of the same color 11

A New Metric: Topical Influence Intuition: the topical influence l (a) of article a is the extent to which it coerces the documents which cite it to have similar topics to it. CitationsInfluence 12

Topical Influence Regression 13 Parameters vector for the Dirichlet prior on the distribution over topics of article a Set of articles that a cites Normalized histogram of topic counts The non-negative scalar topical influence weight for article a

Topical Influence 14 Each article a has a collection of colored balls distributed according to its topic assignments Article aArticle b

Topical Influence 15 Each article a has a collection of colored balls distributed according to its topic assignments It places copies of these balls into the urn for the prior of each article that cites it Article aArticle b Article aArticle b Article cArticle dArticle e

Topical Influence 16 Each article a has a collection of colored balls distributed according to its topic assignments It places copies of these balls into the urn for the prior of each document that cites it Article aArticle b Article aArticle b Article cArticle dArticle e

Topical Influence 17 Each article a has a collection of colored balls distributed according to its topic assignments It places copies of these balls into the urn for the prior of each document that cites it Article aArticle b Article aArticle b Article cArticle dArticle e

Topical Influence 18 Each article a has a collection of colored balls distributed according to its topic assignments It places copies of these balls into the urn for the prior of each document that cites it Article aArticle b Article aArticle b Article cArticle dArticle e

Topical Influence 19 The topical influence weight specifies how many balls article a puts into each citing document’s urn (possibly fractional) l (a) = 5l (b) = 5

Topical Influence 20 l (a) = 10l (b) = 5 The topical influence weight specifies how many balls article a puts into each citing document’s urn (possibly fractional)

Total Topical Influence 21 Total topical influence T (a) is defined to be the total number of balls article a adds to the other articles’ urns T (a) = 20T (b) = 10 l (a) = 10l (b) = 5

Topical Influence Regression for Edge-level Influence Weights 22 We can extend the model to handle differing influence weights on citation edges:

Topical Influence Regression for Edge-level Influence Weights 23 We can extend the model to handle differing influence weights on citation edges:

Inference Collapsed Gibbs sampler Interleave gradient updates for the influence variables (stochastic EM) 24

Inference – Collapsed Gibbs Sampler Usual LDA update, but with topical influence prior 25

Inference – Collapsed Gibbs Sampler Usual LDA update, but with topical influence prior Likelihood for a Polya urn distribution. 26

Experiments Two corpora of scientific articles were used – ACL ( ), 3286 articles – NIPS ( ), 1740 articles – Only citations within the corpora were considered Model validation using metadata Held-out log-likelihood Qualitative analysis 27

Model Validation Using Metadata: Number of times the citation occurs in the text 28

Self citations 29 ACL CorpusNIPS Corpus

Log-Likelihood on Held-Out Documents vs LDA ACLNIPS WinsLossesAverage Improvement WinsLossesAverage Improvement TIR TIRE

Log-Likelihood on Held-Out Documents vs LDA ACLNIPS WinsLossesAverage Improvement WinsLossesAverage Improvement TIR TIRE DMR

Results: Most Influential ACL Articles 32

Results: Most Influential ACL Articles ACL Best Paper Award, 2005 Down to 5 th place, from 1 st by citation count 33

Results: Most Influential NIPS Articles 34

Results: Most Influential NIPS Articles Down to 13 th place, from 1 st by citation count Seminal papers 35

An Optimal-time Binarization Algorithm for Linear Context-Free Rewriting Systems with Fan-out Two. C. Gomez-Rodriguez, G. Satta. Results: Edge Influences, ACL 36 A Hierarchical Phrase-Based Model for Statistical Machine Translation. D. Chiang. Discriminative Training and Maximum Entropy Models for Statistical Machine Translation. F. Och and H. Ney. BLEU: a Method for Automatic Evaluation of Machine Translation. K. Papineni, S. Roukos, T. Ward, W. Zhu. Toward Smaller, Faster, and Better Hierarchical Phrase-based SMT. M. Yang, J. Zheng

An Optimal-time Binarization Algorithm for Linear Context-Free Rewriting Systems with Fan-out Two. C. Gomez-Rodriguez, G. Satta. Results: Edge Influences, ACL 37 A Hierarchical Phrase-Based Model for Statistical Machine Translation. D. Chiang. Discriminative Training and Maximum Entropy Models for Statistical Machine Translation. F. Och and H. Ney. BLEU: a Method for Automatic Evaluation of Machine Translation. K. Papineni, S. Roukos, T. Ward, W. Zhu. Toward Smaller, Faster, and Better Hierarchical Phrase-based SMT. M. Yang, J. Zheng Related SMT paper BLEU evaluation technique Builds upon the method Not related

Multi-time Models for Temporally Abstract Planning. D. Precup, R. Sutton. Results: Edge Influences, NIPS 38 Feudal Reinforcement Learning. P. Dayan, G. Hinton Memory-based Reinforcement Learning: Efficient Computation with Prioritized Sweeping. A. Moore, C. Atkeson. A Delay-Line Based Motion Detection Chip. T. Horiuchi, J. Lazzaro, A. Moore, C. Koch. The Parti-Game Algorithm for Variable Resolution Reinforcement Learning in Multidimensional State-Spaces. A. Moore

Multi-time Models for Temporally Abstract Planning. D. Precup, R. Sutton. Results: Edge Influences, NIPS 39 Feudal Reinforcement Learning. P. Dayan, G. Hinton Memory-based Reinforcement Learning: Efficient Computation with Prioritized Sweeping. A. Moore, C. Atkeson. A Delay-Line Based Motion Detection Chip. T. Horiuchi, J. Lazzaro, A. Moore, C. Koch. The Parti-Game Algorithm for Variable Resolution Reinforcement Learning in Multidimensional State-Spaces. A. Moore Irrelevant Less relevant

Conclusions / Future Work Topical Influence is a quantitative measure of scientific impact which exploits the content of the articles as well as the citation graph Topical Influence Regression can be used to infer topical influence, per article and per citation edge Future work – Authors, journals – Citation context – Temporal dynamics – Application to social media – Other dimensions of scientific importance 40

Thanks! Questions? 41