People Forecasting Where people are going?

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Google News Personalization: Scalable Online Collaborative Filtering
Autonomic Scaling of Cloud Computing Resources
Viral Marketing – Learning Influence Probabilities.
LEARNING INFLUENCE PROBABILITIES IN SOCIAL NETWORKS Amit Goyal Francesco Bonchi Laks V. S. Lakshmanan University of British Columbia Yahoo! Research University.
Online Filtering, Smoothing & Probabilistic Modeling of Streaming Data In short, Applying probabilistic models to Streams Bhargav Kanagal & Amol Deshpande.
Active Learning for Streaming Networked Data Zhilin Yang, Jie Tang, Yutao Zhang Computer Science Department, Tsinghua University.
Junction Trees And Belief Propagation. Junction Trees: Motivation What if we want to compute all marginals, not just one? Doing variable elimination for.
Lirong Xia Approximate inference: Particle filter Tue, April 1, 2014.
Page 1 Hidden Markov Models for Automatic Speech Recognition Dr. Mike Johnson Marquette University, EECE Dept.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Tracking Moving Objects in Anonymized Trajectories Nikolay Vyahhi 1, Spiridon Bakiras 2, Panos Kalnis 3, and Gabriel Ghinita 3 1 St. Petersburg State University.
Learning Bayesian Networks
Probabilistic Analysis of a Large-Scale Urban Traffic Sensor Data Set Jon Hutchins, Alexander Ihler, and Padhraic Smyth Department of Computer Science.
CIS 410/510 Probabilistic Methods for Artificial Intelligence Instructor: Daniel Lowd.
Modeling Count Data over Time Using Dynamic Bayesian Networks Jonathan Hutchins Advisors: Professor Ihler and Professor Smyth.
Efficient Query Evaluation over Temporally Correlated Probabilistic Streams Bhargav Kanagal, Amol Deshpande ΗΥ-562 Advanced Topics on Databases Αλέκα Σεληνιωτάκη.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Charu Aggarwal + * Department of Computer Science, University of Texas at Dallas + IBM T. J. Watson.
1 A Bayesian Method for Guessing the Extreme Values in a Data Set Mingxi Wu, Chris Jermaine University of Florida September 2007.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
Inferring High-Level Behavior from Low-Level Sensors Don Peterson, Lin Liao, Dieter Fox, Henry Kautz Published in UBICOMP 2003 ICS 280.
Direct Message Passing for Hybrid Bayesian Networks Wei Sun, PhD Assistant Research Professor SFL, C4I Center, SEOR Dept. George Mason University, 2009.
Learning Geographical Preferences for Point-of-Interest Recommendation Author(s): Bin Liu Yanjie Fu, Zijun Yao, Hui Xiong [KDD-2013]
Learning and Inferring Transportation Routines By: Lin Liao, Dieter Fox and Henry Kautz Best Paper award AAAI’04.
Dynamic Bayesian Networks and Particle Filtering COMPSCI 276 (chapter 15, Russel and Norvig) 2007.
Exploit of Online Social Networks with Community-Based Graph Semi-Supervised Learning Mingzhen Mo and Irwin King Department of Computer Science and Engineering.
CS Statistical Machine learning Lecture 24
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Presentation Template KwangSoo Yang Florida Atlantic University College of Engineering & Computer Science.
Inferring High-Level Behavior from Low-Level Sensors Donald J. Patterson, Lin Liao, Dieter Fox, and Henry Kautz.
Indexing Correlated Probabilistic Databases Bhargav Kanagal, Amol Deshpande University of Maryland, College Park, USA SIGMOD Presented.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
CPSC 422, Lecture 17Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 17 Oct, 19, 2015 Slide Sources D. Koller, Stanford CS - Probabilistic.
Javad Azimi, Ali Jalali, Xiaoli Fern Oregon State University University of Texas at Austin In NIPS 2011, Workshop in Bayesian optimization, experimental.
Bayes network inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y 
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Learning and Inferring Transportation Routines Lin Liao, Don Patterson, Dieter Fox, Henry Kautz Department of Computer Science and Engineering University.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
CS498-EA Reasoning in AI Lecture #23 Instructor: Eyal Amir Fall Semester 2011.
Hidden Markov Models BMI/CS 576
How to forecast solar flares?
Learning Deep Generative Models by Ruslan Salakhutdinov
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Constrained Hidden Markov Models for Population-based Haplotyping
Boosted Augmented Naive Bayes. Efficient discriminative learning of
A paper on Join Synopses for Approximate Query Answering
Probabilistic Robotics
Modeling and Simulation CS 313
Recovering Temporally Rewiring Networks: A Model-based Approach
Information Based Criteria for Design of Experiments
Objective of This Course
TransCAD Vehicle Routing 2018/11/29.
Particle Filtering ICS 275b 2002.
Discrete Event Simulation - 4
Pattern Recognition and Image Analysis
Assignment, pencil, red pen, highlighter, textbook, GP notebook, calculator Have out: U3D8 Bellwork: Solve for x. Be able to give the exact and approximate.
Scaling up Link Prediction with Ensembles
Bayesian Deep Learning on a Quantum Computer
LECTURE 15: REESTIMATION, EM AND MIXTURES
Chapter14-cont..
Topic Models in Text Processing
False discovery rate estimation
HMM: Particle filters Lirong Xia. HMM: Particle filters Lirong Xia.
Overview: Chapter 2 Localization and Tracking
Universal microbial diagnostics using random DNA probes
Presentation transcript:

People Forecasting Where people are going? Vibhav Gogate-Computer Science Rina Dechter-Computer Science Other Collaborators: Bozhena Bidyuk-Computer Science James Marca-Transportation Science Craig Rindt-Transportation Science University of California, Irvine, CA 92967

Motivation Origin/Destination (O-D) matrix A necessary input to most microscopic simulation models in transportation literature. Old method (Peeta et al. 2002) uses paper and pencil surveys to generate O-D matrices Can IT help? Our proposed “Activity Model” learns and predicts a user’s origins/destinations and routes using his GPS log. Our hope is that even if a small sample of population agrees to share their GPS data, we can compute the aggregate O-D matrix for a large area like a city. So for people who do not know what transportation simulation models are, they are a way to achieve by simulation what you cannot do due to limitations of IT or lack of data. They are of 2 types: Microscopic simulation models, macroscopic simulation models.

How O-D matrices are estimated? (Peeta et al. 2002)

Architecture Learning Engine Probabilistic Model GIS Database O-D matrix for a region Probabilistic Model GIS Database Inference Engine

Probabilistic model: Hybrid Dynamic Mixed Networks Extends Hybrid Dynamic Bayesian Networks (Lerner 2002) to include Discrete constraints Able to model all of the following: Discrete, Continuous Gaussian variables Markov processes Deterministic (constraint networks)+probabilistic information (Bayesian Networks)

Building the activity model dt wt dt-1 wt-1 D: Time-of-day (discrete) W: Day of week (discrete) Goal: collection of locations where the person spends significant amount of time. (discrete) F: Counter to control goal switching. Route: A hidden variable that just predicts what path the person takes (discrete) Location: A pair (e,d) e is the edge on which the person is and d is the distance of the person from one of the end-points of the edge (continuous) Velocity: Continuous GPS reading: (lat,lon,spd,utc). gt-1 gt Ft-1 Ft rt-1 rt vt-1 vt lt-1 lt yt-1 yt

Constraints in the model If (distance(lt-1,gt-1)<=threshold and Ft-1=0) Then Ft=D If (distance(lt-1,gt-1)<=threshold and Ft-1>0) Then Ft=Ft-1-1 If(distance(lt-1,gt-1)>threshold and Ft-1 = 0) Then Ft=0 If(distance(lt-1,gt-1)>threshold and Ft-1 > 0) Then Ft=0 If(Ft-1>0 and Ft=0) gt is given by P(gt|gt-1) If(Ft-1=0 and Ft=0) gt is same as gt-1 If(Ft-1>0 and Ft>0) gt is same as gt-1 If(Ft-1=0 and Ft>0) gt is given by P(gt|gt-1) gt gt-1 Ft-1 Ft lt-1

Example Queries Where the person will be 10 minutes from now? P(lT|d1:t,w1:t,y1:t) where T=t+10 minutes What is the person’s next goal? P(gT|d1:t,w1:t,y1:t)

Example of Goals This is one result of our experiment. In this experiment, a graduate student carries this GPS for three months. We get the GPS log. We put it into our system without any labeling. After a few hours, the system output such a picture, this picture tells us where the common goals are for the person. We then label them manually to make it clear (e.g. Home, church etc).

Example of Route Route Seen Route Predicted Grocery store Once a probabilistic model is learnt, we can use our system to predict what route the person is more likely to take and what is his current goal. For example, the route on the left was seen in the form of GPS reading by our model and the route on the right was predicted by our model. As we can see our model predicted that the person was likely to go to a grocery store. Grocery store Route Seen Route Predicted

Contributions A new modeling framework of “Hybrid Dynamic Mixed Networks” A “Hybrid Dynamic Mixed Network model” for transportation routines Predict origin/destinations and routes taken by an individual. Novel inference algorithms for reasoning in Hybrid Dynamic Mixed Networks An Expectation propagation based algorithm A new algorithm that combines Particle Filtering and Generalized Belief Propagation in a systematic way.

Inference in Hybrid Dynamic Mixed Networks (HDMN) Filtering problem: The Belief state at time t given evidence until time t : P(Xt|e1:t) Complexity of exact inference: NP-hard Exponential in treewidth Discrete Dynamic Mixed Networks Treewidth = number of variables in each time slice Hybrid Dynamic Mixed Networks Treewidth = O(T) where T is the number of time-slices. Approximation is a must in most cases!

Approximate Inference Two popular approximate inference algorithms for Dynamic Networks Generalized Belief Propagation (Heskes et al. ’02) Rao-Blackwellised Particle Filtering (Doucet et al. ’02) Our contribution: Extend these two algorithms to allow discrete constraints Iterative Join Graph Propagation-Sequential (IJGP-S) A new Rao-Blackwellised Particle Filtering (RBPF) algorithm called IJGP-RBPF. Use output of IJGP to compute an importance function Parameterized by two complexity parameters of “i” and “w” which provides us with a range of algorithms to choose from.

Experimental Results: Data Collection GPS data was collected by one of the authors for a period of 6 months. Latitude and longitude pairs 3 months data was used for training and 3 months for testing. Data divided into segments A segment is a series of GPS readings such that two consecutive readings are less than 15 minutes apart.

Experimental Results: Models and algorithms Test if adding new variables improves prediction accuracy. Model-1: Model as described before Model-2: Remove variables dt and wt Model-3: Remove variables dt, wt,ft,rt,gt from each time slice. Algorithms: IJGP-RBPF(1,2), IJGP-RBPF(2,1), IJGP-S(1) and IJGP-S(2)

Various Activity models dt wt dt-1 wt-1 gt-1 gt Model-1 Ft-1 Ft rt-1 rt Model-2 vt-1 vt Model-3 lt-1 lt yt-1 yt

Learning the models from data EM algorithm used for learning the models Takes about 3 to 5 days to learn data that is distributed over 3 months. Since EM uses inference as a sub-step, we have 4 EM algorithms corresponding to the 4 algorithms used for inference IJGP-RBPF(1,2), IJGP-RBPF(2,1), IJGP-S(1) and IJGP-S(2)

Predicting Goals (MODEL-1) Compute P(gt|e1:t) and compare it with the actual goal. Accuracy = percentage of goals predicted correctly. N = number of particles Column: learning algorithm Row: inference algorithm

Predicting Goals (Model-2) Compute P(gt|e1:t) and compare it with the actual goal. Accuracy = percentage of goals predicted correctly. N = number of particles Column: learning algorithm Row: inference algorithm

Predicting Goals (Model-3) Compute P(gt|e1:t) and compare it with the actual goal. Accuracy = percentage of goals predicted correctly. N = number of particles Column: learning algorithm Row: inference algorithm

Predicting Routes Compare the path of the person predicted by the model with the actual path. False positives (FP)---Precision count the number of roads that were not taken by the person but were in the predicted path. False Negatives (FN)---Recall count the number of roads that were taken by the person but were not in the predicted path.

False Positives and False Negatives for Route prediction Model-1 shows the highest route prediction accuracy, given by low false positives and false negatives.

Future Work: O-D estimation through Simulation Randomly generate regions and a population Land-use structures through Microsoft Map-point Assume an activity model per individual Simulation gives an aggregate O-D matrix (called actual O-D) Take a random sample of the population and use their GPS data Our proposed System would predict an O-D matrix (predicted O-D) Success: Distance between predicted and accurate O-D matrix.

Challenges Scalable algorithms Our algorithms take about 3-5 days/individual! Does the proposed simulation represent real-world? A model for data aggregation Inference and Learning for other continuous frameworks Poisson distribution. Discrete children of continuous parents.