Extracting Mobile Behavioral Patterns with the Distant N-Gram Topic Model Lingzi Hong Feb 10th.

Slides:



Advertisements
Similar presentations
Topic models Source: Topic models, David Blei, MLSS 09.
Advertisements

An Adaptive Learning Method for Target Tracking across Multiple Cameras Kuan-Wen Chen, Chih-Chuan Lai, Yi-Ping Hung, Chu-Song Chen National Taiwan University.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Ouyang Ruofei Topic Model Latent Dirichlet Allocation Ouyang Ruofei May LDA.
Hierarchical Dirichlet Processes
Bayesian Estimation in MARK
Title: The Author-Topic Model for Authors and Documents
Patch to the Future: Unsupervised Visual Prediction
Generative Topic Models for Community Analysis
Caimei Lu et al. (KDD 2010) Presented by Anson Liang.
HMM-BASED PATTERN DETECTION. Outline  Markov Process  Hidden Markov Models Elements Basic Problems Evaluation Optimization Training Implementation 2-D.
Unsupervised discovery of visual object class hierarchies Josef Sivic (INRIA / ENS), Bryan Russell (MIT), Andrew Zisserman (Oxford), Alyosha Efros (CMU)
Generative learning methods for bags of features
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Scalable Text Mining with Sparse Generative Models
Correlated Topic Models By Blei and Lafferty (NIPS 2005) Presented by Chunping Wang ECE, Duke University August 4 th, 2006.
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
1 Linmei HU 1, Juanzi LI 1, Zhihui LI 2, Chao SHAO 1, and Zhixing LI 1 1 Knowledge Engineering Group, Dept. of Computer Science and Technology, Tsinghua.
Topic Modelling: Beyond Bag of Words By Hanna M. Wallach ICML 2006 Presented by Eric Wang, April 25 th 2008.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Hidden Topic Markov Models Amit Gruber, Michal Rosen-Zvi and Yair Weiss in AISTATS 2007 Discussion led by Chunping Wang ECE, Duke University March 2, 2009.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Integrating Topics and Syntax -Thomas L
Summary We propose a framework for jointly modeling networks and text associated with them, such as networks or user review websites. The proposed.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Latent Dirichlet Allocation D. Blei, A. Ng, and M. Jordan. Journal of Machine Learning Research, 3: , January Jonathan Huang
Probabilistic Models for Discovering E-Communities Ding Zhou, Eren Manavoglu, Jia Li, C. Lee Giles, Hongyuan Zha The Pennsylvania State University WWW.
Introduction to LDA Jinyang Gao. Outline Bayesian Analysis Dirichlet Distribution Evolution of Topic Model Gibbs Sampling Intuition Analysis of Parameter.
National Taiwan University, Taiwan
Topic Modeling using Latent Dirichlet Allocation
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
CS246 Latent Dirichlet Analysis. LSI  LSI uses SVD to find the best rank-K approximation  The result is difficult to interpret especially with negative.
Dynamic Multi-Faceted Topic Discovery in Twitter Date : 2013/11/27 Source : CIKM’13 Advisor : Dr.Jia-ling, Koh Speaker : Wei, Chang 1.
Multi-target Detection in Sensor Networks Xiaoling Wang ECE691, Fall 2003.
Hidden Variables, the EM Algorithm, and Mixtures of Gaussians Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem 02/22/11.
A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation Yee W. Teh, David Newman and Max Welling Published on NIPS 2006 Discussion.
Understanding unstructured texts via Latent Dirichlet Allocation Raphael Cohen DSaaS, EMC IT June 2015.
TribeFlow Mining & Predicting User Trajectories Flavio Figueiredo Bruno Ribeiro Jussara M. AlmeidaChristos Faloutsos 1.
Chapter 6 Activity Recognition from Trajectory Data Yin Zhu, Vincent Zheng and Qiang Yang HKUST November 2011.
Canadian Bioinformatics Workshops
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Fast search for Dirichlet process mixture models
Topic Modeling for Short Texts with Auxiliary Word Embeddings
Kuifei Yu, Baoxian Zhang, Hengshu Zhu,Huanhuan Cao, and Jilei Tian
Machine Learning and Data Mining Clustering
WP2 INERTIA Distributed Multi-Agent Based Framework
CSRS: A Context and Sequence Aware Recommendation System
Online Multiscale Dynamic Topic Models
Shuang-Hong Yang, Hongyuan Zha, Bao-Gang Hu NIPS2009
Entity- & Topic-Based Information Ordering
Multimodal Learning with Deep Boltzmann Machines
Trevor Savage, Bogdan Dit, Malcom Gethers and Denys Poshyvanyk
Latent Dirichlet Analysis
People-LDA using Face Recognition
Multidimensional Integration Part I
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
The topic discovery models
Topic Modeling Nick Jordan.
Bayesian Inference for Mixture Language Models
Stochastic Optimization Maximization for Latent Variable Models
iSRD Spam Review Detection with Imbalanced Data Distributions
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
Topic models for corpora and for graphs
Michal Rosen-Zvi University of California, Irvine
Junghoo “John” Cho UCLA
Topic models for corpora and for graphs
Topic Models in Text Processing
Machine Learning and Data Mining Clustering
INF 141: Information Retrieval
Conceptual grounding Nisheeth 26th March 2019.
Presentation transcript:

Extracting Mobile Behavioral Patterns with the Distant N-Gram Topic Model Lingzi Hong Feb 10th

Research Question problem: modeling activity sequences for large-scale human routine discovery from cellphone censor data fundamental difficulties: do not know the basic units of time for the activities in the question. (hourly,daily?) =>effective modeling of multiple unknown time duration

focus on Probabilistic Topic Models unsupervised=>mining structure of data handle uncertainty extended in various ways to integrate multiple data types=>sensor activity sequences

contributions propose the distant n-gram topic model (DNTM) for sequence modeling derive inference process using Markov Chain Monte Carlo (MCMC) sampling apply to two real large-scale datasets comparative analysis with Latent Dirichlet Allocation (LDA)

Related Work Topic model as a useful tool 1. T. Huynh, M. Fritz, and B. Schiele. Discovery of activity patterns using topic models. 2. K. Farrahi and D. Gatica-Perez. Probabilistic mining of socio- geographic routines from mobile phone data. 3. T. Bao, H. Cao, E. Chen, J. Tian, and H. Xiong. An unsupervised approach to modeling personalized contexts of mobile users. 4. K. Farrahi and D. Gatica-Perez. Discovering routines from large-scale human locations using probabilistic topic models. Topic model in terms of text 1. LDA. determine probability of each word to each topic and probability of each topic given each document N-gram discovery 1. bigram topic model 2. topic n-gram model

Distant N-Gram Topic Model q m corpus Sm w1,w2,…,wN w = (t, l) t-location l-coordinate of a day The distribution of W1 given topics

Distant N-Gram Topic Model General process: 1. Initialization (document topic, distribution over labels) 2. Sequence generation procedure (estimate paratemeters) model parameters derived based on MCMC approach of Gibbs sampling estimation of parameters: ?

Distant N-Gram Topic Model Anyway there is code that helps to implement this process

Experiments and Results Nokia Smartphone Data Tricks?: days with topic distribution => 10 most probable days for the topic ranked from top to bottom

Experiments and Results MIT Reality Mining Data L={‘H’,’W’,’O’,’N’}, tt=48 most probable days given topics

Experiments and Results most probable sequence components for topics

Evaluation splitting into training and testing log-likihood: A test set is a collection of unseen documents wd, the model is described by the topic matrix Φ, and the hyperparameter α for topic-distribution of documents. log-likihood: The probability of unseen held-out documents given some training documents. Higher likelihood implies a better model Perplexity: The lower perplexity the better the model

Evaluation perplexity of the DNTM over number of 20% unseen days Average log-likelihood of the DNTM versus LDA on 20% unseen days.

Discussion generalization of the model model assumes every topic has a distribution of sequence q, with element w labeled with time and location, which means w involves with a general topic distribution. But if there is a lot of user samples, a workplace for A might be leisure place for B. For topic models, if one word involves with a topic distribution, this distribution will be equally applied to all documents. However we can’t assume a place has the same topic distribution of day activities for different people. Could we? Nokia Smartphone: 2 users and each with a lot of places in two different cities. Few overlapping places with mixed function. Result is separately for user1 and user2. MIT data: lots of users but places have been labeled. So result is only identification of topics. Real data set will include a lot of users and not labeled places.

Discussion How to choose N? Segmentation of sequences according to activities or according to time? What if the last sequence q is not complete?

Discussion Could we just make clustering of the sequences to detect activity patterns? 48 intervals a day, each interval as a feature, value of the feature is the label (‘H’,’W’,’O’,’N’)