Human-centered Machine Learning

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

TOPDRIM: Update WP2 March 2013 Rick Quax, Peter M.A. Sloot.
Unità di Perugia e di Roma “Tor Vergata” "Uncertain production systems: optimal feedback control of the single site and extension to the multi-site case"
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
Statistical Probabilistic Model Checking Håkan L. S. Younes Carnegie Mellon University.
Part 4 b Forward-Backward Algorithm & Viterbi Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Probability theory 2010 Conditional distributions  Conditional probability:  Conditional probability mass function: Discrete case  Conditional probability.
Space-time Modelling Using Differential Equations Alan E. Gelfand, ISDS, Duke University (with J. Duan and G. Puggioni)
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Topic models for corpora and for graphs. Motivation Social graphs seem to have –some aspects of randomness small diameter, giant connected components,..
Theory of Decision Time Dynamics, with Applications to Memory.
Asymptotic Techniques
Modeling Your Spiking Data with Generalized Linear Models.
UIUC CS 498: Section EA Lecture #21 Reasoning in Artificial Intelligence Professor: Eyal Amir Fall Semester 2011 (Some slides from Kevin Murphy (UBC))
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
TheoryApplication Discrete Continuous - - Stochastic differential equations - - Ito’s formula - - Derivation of the Black-Scholes equation - - Markov processes.
Generalized Semi- Markov Processes (GSMP). Summary Some Definitions The Poisson Process Properties of the Poisson Process  Interarrival times  Memoryless.
ECE 466/658: Performance Evaluation and Simulation Introduction Instructor: Christos Panayiotou.
Lecture 2: Statistical learning primer for biologists
ECE 8443 – Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem Proof EM Example – Missing Data Intro to Hidden Markov Models.
Manuel Gomez Rodriguez Bernhard Schölkopf I NFLUENCE M AXIMIZATION IN C ONTINUOUS T IME D IFFUSION N ETWORKS , ICML ‘12.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Jensen’s Inequality (Special Case) EM Theorem.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
A Method to Approximate the Bayesian Posterior Distribution in Singular Learning Machines Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
CS498-EA Reasoning in AI Lecture #23 Instructor: Eyal Amir Fall Semester 2011.
Machine learning for Dynamic Social Network Analysis
OPERATING SYSTEMS CS 3502 Fall 2017
Machine learning for Dynamic Social Network Analysis
Manuel Gomez Rodriguez
Learning Deep Generative Models by Ruslan Salakhutdinov
Availability Availability - A(t)
Machine learning for Dynamic Social Network Analysis
Probability Theory and Parameter Estimation I
Bayesian estimation Bayes’s theorem: prior, likelihood, posterior
Tracking We are given a contour G1 with coordinates G1={x1 , x2 , … , xN} at the initial frame t=1, were the image is It=1 . We are interested in tracking.
Industrial Engineering Dep
Models.
Simulation Statistics
Manuel Gomez Rodriguez
Chapter 6: Temporal Difference Learning
LECTURE 10: EXPECTATION MAXIMIZATION (EM)
Probabilistic Robotics
Chapter 7: Sampling Distributions
A Latent Space Approach to Dynamic Embedding of Co-occurrence Data
Machine Learning Today: Reading: Maria Florina Balcan
Hidden Markov Models Part 2: Algorithms
Bayesian Models in Machine Learning
Modelling data and curve fitting
TexPoint fonts used in EMF.
Multidimensional Integration Part I
Filtering and State Estimation: Basic Concepts
STOCHASTIC HYDROLOGY Random Processes
Topic models for corpora and for graphs
Chapter 6: Temporal Difference Learning
Topic models for corpora and for graphs
Hydrology Modeling in Alaska: Modeling Overview
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
copyright Robert J. Marks II
CISE-301: Numerical Methods Topic 1: Introduction to Numerical Methods and Taylor Series Lectures 1-4: KFUPM CISE301_Topic1.
1.1 Dynamical Systems MODELING CHANGE
Human-centered Machine Learning
Exploiting Unintended Feature Leakage in Collaborative Learning
Presentation transcript:

Human-centered Machine Learning Opinion Dynamics with Marked Temporal Point Processes (TPPs) Human-centered Machine Learning http://courses.mpi-sws.org/hcml-ws18/

Opinions in social media

Use social media to sense opinions People’s opinion about political discourse Investors’ sentiment about stocks Brand sentiment and reputation

What about opinion dynamics? Complex stochastic processes (often over a network)

Example of opinion dynamics Christine means D follows S Bob Beth Beth is influential Joe David Expressed opinions

Model of opinion dynamics Can we design a realistic model that fits real fine-grained opinion traces? Why this goal? Predict (infer) opinions, even if not expressed! Identify opinion leaders Sergey Bilak

Traditional models of opinion dynamics There are a lot of theoretical models of opinion dynamics, but… 1. Do not distinguish between latent and expressed opinions 2. Opinions are updated sequentially in discrete time 3. Difficult to learn from fine-grained data and thus inaccurate predictions 4. Focus on steady state, neglecting transient behavior

Key ideas of Marked TPP model Alice’s expressed opinions Latent opinions vs expressed opinions t Alice’s latent opinion t Informational and social influence Bob and Charly’s expressed opinions t Bob Charly Alice’s expressed opinions Alice t [De et al., NIPS 2016]

Noisy observation of latent opinion Message representation We represent messages using marked temporal point processes: NA(t) NB(t) NC (t) Message: User Time Sentiment (mark) Noisy observation of latent opinion [De et al., NIPS 2016]

Message intensity Hawkes process User’s intensity NA(t) NB(t) NC (t) Memory Hawkes process User’s intensity Messages on her own initiative Previous messages by user v Influence from user v on user u [De et al., NIPS 2016]

Sentiment distribution Latent opinion Sentiment: Continuous (based on sentiment analysis): It depends on the recorded data Discrete (based on upvotes/downvotes): [De et al., NIPS 2016]

Stochastic process for (latent) opinions Memory User’s latent opinion User’s initial opinion influence from user v on user u Previous sentiment by user v xAlice(t) <0, disagree Bob Christine >0, agree Alice aChristine,Alicem2 aBob,Alicem1 αAlice m3 m(t) m1 m2 [De et al., NIPS 2016]

Stubborness, conformity, and compromise The model allows for: Stubborn users Compromised users Conforming users

Example: positive opinions win t=T t=∞ Average opinion Opinions per node

Example: negative opinions win t=T t=∞ Average opinion Opinions per node

Example: opinions get polarized t=T t=∞ Average opinion Opinions per node

Model inference from opinion data Events likelihood Message sentiments Message times (marks) Theorem. The maximum likelihood problem is convex in the model parameters. Markov property Sums and integrals in linear time! [De et al., NIPS 2016]

Informational influence Opinion model as Jump SDEs Proposition. The tuple (x*(t), λ*(t), N(t)) is a Markov process, whose dynamics are defined by the following marked jumped stochastic differential equations (SDEs) Network! Latent opinions Expressed opinions Informational influence t Message intensities Temporal influence Network! [De et al., NIPS 2016]

Proof sketch of the proposition (I) Let’s do it for one-dimensional Hawkes: t

Proof sketch of the proposition (II)

Opinion forecasting For forecasting, we compute conditional averages: History up to t0 Sources of Randomness If t analytical solution (Th. 2) numerical solution (Th. 4) Otherwise: Sampling based solution

Opinion forecasting The Avengers: Age of Ultron, 05/2015 The forecasted opinion becomes less accurate as T increases, as one may expect. [De et al., NIPS 2016]