F EATURE -E NHANCED P ROBABILISTIC M ODELS FOR D IFFUSION N ETWORK I NFERENCE Stefano Ermon ECML-PKDD September 26, 2012 Joint work with Liaoruo Wang and.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Viral Marketing – Learning Influence Probabilities.
LEARNING INFLUENCE PROBABILITIES IN SOCIAL NETWORKS Amit Goyal Francesco Bonchi Laks V. S. Lakshmanan University of British Columbia Yahoo! Research University.
Active Learning for Streaming Networked Data Zhilin Yang, Jie Tang, Yutao Zhang Computer Science Department, Tsinghua University.
Caroline Rougier, Jean Meunier, Alain St-Arnaud, and Jacqueline Rousseau IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 21, NO. 5,
Algorithms + L. Grewe.
Community Detection Algorithm and Community Quality Metric Mingming Chen & Boleslaw K. Szymanski Department of Computer Science Rensselaer Polytechnic.
Maximizing the Spread of Influence through a Social Network
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Fast Iterative Alignment of Pose Graphs with Poor Initial Estimates Edwin Olson John Leonard, Seth Teller
Hidden Markov Models in NLP
Visual Recognition Tutorial
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
INFERRING NETWORKS OF DIFFUSION AND INFLUENCE Presented by Alicia Frame Paper by Manuel Gomez-Rodriguez, Jure Leskovec, and Andreas Kraus.
Clustering Social Networks Isabelle Stanton, University of Virginia Joint work with Nina Mishra, Robert Schreiber, and Robert E. Tarjan.
1 Localization Technologies for Sensor Networks Craig Gotsman, Technion/Harvard Collaboration with: Yehuda Koren, AT&T Labs.
Decision Trees with Minimal Costs (ICML 2004, Banff, Canada) Charles X. Ling, Univ of Western Ontario, Canada Qiang Yang, HK UST, Hong Kong Jianning Wang,
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Dynamic Programming Introduction to Algorithms Dynamic Programming CSE 680 Prof. Roger Crawfis.
Latent Boosting for Action Recognition Zhi Feng Huang et al. BMVC Jeany Son.
Models of Influence in Online Social Networks
On Ranking and Influence in Social Networks Huy Nguyen Lab seminar November 2, 2012.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Modeling Information Diffusion in Networks with Unobserved Links Quang Duong Michael P. Wellman Satinder Singh Computer Science and Engineering University.
1 1 MPI for Intelligent Systems 2 Stanford University Manuel Gomez Rodriguez 1,2 David Balduzzi 1 Bernhard Schölkopf 1 UNCOVERING THE TEMPORAL DYNAMICS.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Graphical models for part of speech tagging
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
1 1 Stanford University 2 MPI for Biological Cybernetics 3 California Institute of Technology Inferring Networks of Diffusion and Influence Manuel Gomez.
1 1 Stanford University 2 MPI for Biological Cybernetics 3 California Institute of Technology Inferring Networks of Diffusion and Influence Manuel Gomez.
Information Spread and Information Maximization in Social Networks Xie Yiran 5.28.
Bayesian Learning Chapter Some material adapted from lecture notes by Lise Getoor and Ron Parr.
Hidden Markov Models in Keystroke Dynamics Md Liakat Ali, John V. Monaco, and Charles C. Tappert Seidenberg School of CSIS, Pace University, White Plains,
Jing (Selena) He and Hisham M. Haddad Department of Computer Science, Kennesaw State University Shouling Ji, Xiaojing Liao, and Raheem Beyah School of.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
A data-mining approach for multiple structural alignment of proteins WY Siu, N Mamoulis, SM Yiu, HL Chan The University of Hong Kong Sep 9, 2009.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Paired Sampling in Density-Sensitive Active Learning Pinar Donmez joint work with Jaime G. Carbonell Language Technologies Institute School of Computer.
Manuel Gomez Rodriguez Structure and Dynamics of Information Pathways in On-line Media W ORKSHOP M ENORCA, MPI FOR I NTELLIGENT S YSTEMS.
NP-COMPLETE PROBLEMS. Admin  Two more assignments…  No office hours on tomorrow.
Online Social Networks and Media
I NFORMATION C ASCADE Priyanka Garg. OUTLINE Information Propagation Virus Propagation Model How to model infection? Inferring Latent Social Networks.
CS B 553: A LGORITHMS FOR O PTIMIZATION AND L EARNING Parameter Learning with Hidden Variables & Expectation Maximization.
Manuel Gomez Rodriguez Bernhard Schölkopf I NFLUENCE M AXIMIZATION IN C ONTINUOUS T IME D IFFUSION N ETWORKS , ICML ‘12.
Towards Social User Profiling: Unified and Discriminative Influence Model for Inferring Home Locations Rui Li, Shengjie Wang, Hongbo Deng, Rui Wang, Kevin.
DM-MEETING Bijaya Adhikari OUTLINE From Micro to Macro: Uncovering and Predicting Information Cascading Process with Behavioral Dynamics 
The Visual Causality Analyst: An Interactive Interface for Causal Reasoning Jun Wang, Stony Brook University Klaus Mueller, Stony Brook University, SUNY.
DM GROUP MEETING PRESENTATION PLAN Eigenvector-based Centrality Measures For Temporal Networks by D Taylor et.al. Uncovering the Small Community.
A Latent Social Approach to YouTube Popularity Prediction Amandianeze Nwana Prof. Salman Avestimehr Prof. Tsuhan Chen.
1 1 MPI for Intelligent Systems 2 Stanford University Manuel Gomez Rodriguez 1,2 Bernhard Schölkopf 1 S UBMODULAR I NFERENCE OF D IFFUSION NETWORKS FROM.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
1 1 Stanford University 2 MPI for Biological Cybernetics 3 California Institute of Technology Inferring Networks of Diffusion and Influence Manuel Gomez.
Biao Wang 1, Ge Chen 1, Luoyi Fu 1, Li Song 1, Xinbing Wang 1, Xue Liu 2 1 Shanghai Jiao Tong University 2 McGill University
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Inferring Networks of Diffusion and Influence
Independent Cascade Model and Linear Threshold Model
Greedy & Heuristic algorithms in Influence Maximization
Intelligent Information System Lab
Machine Learning Week 10.
Independent Cascade Model and Linear Threshold Model
Learning with information of features
Mixture of Mutually Exciting Processes for Viral Diffusion
Learning Emoji Embeddings Using Emoji Co-Occurrence Network Graph
Scaling up Link Prediction with Ensembles
Learning Probabilistic Graphical Models Overview Learning Problems.
Independent Cascade Model and Linear Threshold Model
Presentation transcript:

F EATURE -E NHANCED P ROBABILISTIC M ODELS FOR D IFFUSION N ETWORK I NFERENCE Stefano Ermon ECML-PKDD September 26, 2012 Joint work with Liaoruo Wang and John E. Hopcroft

B ACKGROUND Diffusion processes common in many types of networks Cascading examples contact networks <> infections friendship networks <> gossips social networks <> products academic networks <> ideas

B ACKGROUND Typically, network structure assumed known Many interesting questions minimize spread (vaccinations) maximize spread (viral marketing) interdictions What if the underlying network is unknown?

N ETWORK I NFERENCE N ET I NF [Gomez-Rodriguez et al. 2010] input: actual number of edges in the latent network observations of information cascades output: set of edges maximizing the likelihood of the observations submodular N ET R ATE [Gomez-Rodriguez et al. 2011] input: observations of information cascades output: set of transmission rates maximizing the likelihood of the observations convex optimization problem

C ASCADES (v 0,t 0 1 ) (v 1,t 1 1 )(v 2,t 2 1 )(v 3,t 3 1 )(v 4,t 4 1 ) (v 5,t 0 2 )(v 6,t 1 2 ) (v 2,t 2 2 ) (v 7,t 3 2 )(v 8,t 4 2 ) (v 9,t 0 3 ) (v 3,t 1 3 ) (v 7,t 2 3 ) (v 10,t 3 3 )(v 11,t 4 3 ) Given observations of a diffusion process, what can we infer about the underlying network?

M OTIVATING E XAMPLE information diffusion in the Twitter following network

P REVIOUS W ORK Major assumptions the diffusion process is causal (not affected by events in the future) the diffusion process is monotonic (can be infected at most once) infection events closer in time are more likely to be causally related (e.g., exponential, Rayleigh, or power-law distribution) Time-stamps are not sufficient most real-world diffusion processes are recurrent cascades are often a mixture of (geographically) local sub-cascades cannot tell them apart by just looking at time-stamps many other informative factors (e.g., language, pairwise similarity) Our work generalizes previous models to take these factors into account.

P ROBLEM D EFINITION Weighted, directed graph G=(V, E) known: node set V unknown: weighted edge set E Observations: generalized cascades {π 1, π 2,…, π M } 957BenFM #ladygaga always rocks… 2frog #ladygaga bella canzone… AbbeyResort #followfriday see you all tonight… figmentations #followfriday cannot wait… 2frog #followfriday 周五活动计划 …

P ROBLEM D EFINITION Given set of vertices V set of generalized cascades {π 1, π 2,…, π M } generative probabilistic model (feature-enhanced) Goal: find the most likely adjacency matrix of transmission rates A={α jk |j,k  V,j  k} latent network observed cascades {π 1, π 2,…, π M }

F EATURE -E NHANCED M ODEL Multiple occurrences splitting: an infection event of a node is the result of all previous events up to its last infection (memoryless) non-splitting: an infection event is the result of all previous events Independent of future infection events (causal process)

F EATURE -E NHANCED M ODEL Generalized cascade: assumption 1: events closer in time are more likely to be causally related assumption 2: events closer in feature space are more likely to be causally related

G ENERATIVE M ODEL diffusion distribution (exponental, Rayleigh, etc.) assumed network A Given model and observed cascades, the likelihood of an assumed network A : enough edges so that every infection event can be explained (reward) for every infected node, for each of its neighbors, how long does it take for the neighbor to become infected? (penalty) why not infected at all? (penalty) distance between events probability of being causally related

O PTIMIZATION F RAMEWORK assumed network A maximize L ( π 1, π 2 | A ) 1. convex in A 2. decomposable diffusion distribution (exponental, Rayleigh, etc.)

E XPERIMENTAL S ETUP Dataset Twitter (66,679 nodes; 240,637 directed edges) Cascades (500 hashtags; 103,148 tweets) Ground truth known Feature Model language pairwise similarity combination

E XPERIMENTAL S ETUP Baselines N ET I NF (takes true number of edges as input) N ET R ATE Language Detector the language is computed using the n-gram model noisy estimates Convex Optimization limited-memory BFGS algorithm with box constraints CVXOPT cannot handle the scale of our Twitter dataset All algorithms are implemented using Python with the Fortran implementation of LBFGS-B available in Scipy, and all experiments are performed on a machine running CentOS Linux with a 6-core Intel x GHZ CPU and 48GB memory.

P ERFORMANCE C OMPARISON Non-Splitting Exponential M ETRIC N ET I NF N ET R ATEMO N ETMO N ET +L MO N ET +J MO N ET +LJ P RECISION R ECALL F1- SCORE TP FP FN %

P ERFORMANCE C OMPARISON Splitting Exponential M ETRIC N ET I NF N ET R ATEMO N ETMO N ET +L MO N ET +J MO N ET +LJ P RECISION R ECALL F1- SCORE TP FP FN %

P ERFORMANCE C OMPARISON Non-Splitting Rayleigh M ETRIC N ET I NF N ET R ATEMO N ETMO N ET +L MO N ET +J MO N ET +LJ P RECISION R ECALL F1- SCORE TP FP FN %

P ERFORMANCE C OMPARISON Splitting Rayleigh M ETRIC N ET I NF N ET R ATEMO N ETMO N ET +L MO N ET +J MO N ET +LJ P RECISION R ECALL F1- SCORE TP FP FN %

C ONCLUSION Feature-enhanced probabilistic models to infer the latent network from observations of a diffusion process Primary approach MO N ET with non-splitting and splitting solutions to handle recurrent processes Our models consider not only the relative time differences between infection events, but also a richer set of features. The inference problem still involves convex optimization. It can be decomposed into smaller sub-problems that we can efficiently solve in parallel. Improved performance on Twitter

THANK YOU!