Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields 2012311529 Yong-Joong Kim Dept. of Computer Science Yonsei.

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Yinyin Yuan and Chang-Tsun Li Computer Science Department
Learning on the Test Data: Leveraging “Unseen” Features Ben Taskar Ming FaiWong Daphne Koller.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
Constrained Approximate Maximum Entropy Learning (CAMEL) Varun Ganapathi, David Vickrey, John Duchi, Daphne Koller Stanford University TexPoint fonts used.
Exact Inference in Bayes Nets
Carolina Galleguillos, Brian McFee, Serge Belongie, Gert Lanckriet Computer Science and Engineering Department Electrical and Computer Engineering Department.
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data John Lafferty Andrew McCallum Fernando Pereira.
Markov Networks.
Hidden Markov Models Theory By Johan Walters (SR 2003)
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
A Graphical Model For Simultaneous Partitioning And Labeling Philip Cowans & Martin Szummer AISTATS, Jan 2005 Cambridge.
Overview Full Bayesian Learning MAP learning
Learning to Detect A Salient Object Reporter: 鄭綱 (3/2)
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Practical Belief Propagation in Wireless Sensor Networks Bracha Hod Based on a joint work with: Danny Dolev, Tal Anker and Danny Bickson The Hebrew University.
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Lecture 5: Learning models using EM
Abstract We present a model of curvilinear grouping using piecewise linear representations of contours and a conditional random field to capture continuity.
Conditional Random Fields
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Abstract Extracting a matte by previous approaches require the input image to be pre-segmented into three regions (trimap). This pre-segmentation based.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Belief Propagation Kai Ju Liu March 9, Statistical Problems Medicine Finance Internet Computer vision.
Cue Integration in Figure/Ground Labeling Xiaofeng Ren, Charless Fowlkes and Jitendra Malik, U.C. Berkeley We present a model of edge and region grouping.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Computer vision: models, learning and inference
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Charu Aggarwal + * Department of Computer Science, University of Texas at Dallas + IBM T. J. Watson.
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren.
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Inferring High-Level Behavior from Low-Level Sensors Don Peterson, Lin Liao, Dieter Fox, Henry Kautz Published in UBICOMP 2003 ICS 280.
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Multimodal Information Analysis for Emotion Recognition
Learning and Inferring Transportation Routines By: Lin Liao, Dieter Fox and Henry Kautz Best Paper award AAAI’04.
28 February, 2003University of Glasgow1 Cluster Variation Method and Probabilistic Image Processing -- Loopy Belief Propagation -- Kazuyuki Tanaka Graduate.
Continuous Variables Write message update equation as an expectation: Proposal distribution W t (x t ) for each node Samples define a random discretization.
An Asymptotic Analysis of Generative, Discriminative, and Pseudolikelihood Estimators by Percy Liang and Michael Jordan (ICML 2008 ) Presented by Lihan.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Reestimation Equations Continuous Distributions.
Presented by Jian-Shiun Tzeng 5/7/2009 Conditional Random Fields: An Introduction Hanna M. Wallach University of Pennsylvania CIS Technical Report MS-CIS
Interactive Learning of the Acoustic Properties of Objects by a Robot
Tell Me What You See and I will Show You Where It Is Jia Xu 1 Alexander G. Schwing 2 Raquel Urtasun 2,3 1 University of Wisconsin-Madison 2 University.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Training Conditional Random Fields using Virtual Evidence Boosting Lin Liao, Tanzeem Choudhury †, Dieter Fox, and Henry Kautz University of Washington.
Joseph Xu Soar Workshop Learning Modal Continuous Models.
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
Inferring High-Level Behavior from Low-Level Sensors Donald J. Patterson, Lin Liao, Dieter Fox, and Henry Kautz.
Unsupervised Mining of Statistical Temporal Structures in Video Liu ze yuan May 15,2011.
Indexing Correlated Probabilistic Databases Bhargav Kanagal, Amol Deshpande University of Maryland, College Park, USA SIGMOD Presented.
John Lafferty Andrew McCallum Fernando Pereira
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
Objectives: Terminology Components The Design Cycle Resources: DHS Slides – Chapter 1 Glossary Java Applet URL:.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.
Maximum Entropy Model, Bayesian Networks, HMM, Markov Random Fields, (Hidden/Segmental) Conditional Random Fields.
Markov Random Fields & Conditional Random Fields
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Pattern Recognition and Machine Learning
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Statistical Models for Automatic Speech Recognition Lukáš Burget.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Hidden Markov Model Parameter Estimation BMI/CS 576 Colin Dewey Fall 2015.
Distributed cooperation and coordination using the Max-Sum algorithm
SA-1 University of Washington Department of Computer Science & Engineering Robotics and State Estimation Lab Dieter Fox Stephen Friedman, Lin Liao, Benson.
Thrust IIA: Environmental State Estimation and Mapping Dieter Fox (Lead) Nicholas Roy MURI 8 Kickoff Meeting 2007.
Graphical Models for Segmenting and Labeling Sequence Data Manoj Kumar Chinnakotla NLP-AI Seminar.
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Expectation-Maximization & Belief Propagation
Markov Networks.
Presentation transcript:

Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei University Lin Liao, Dieter Fox, and Henry Kautz, In International Journal of Robotics Research (IJRR), 26(1), 2007

Contents Motivation Hierarchical Activity Model Preliminaries : Conditional Random Fields – Overview – Inference – Parameter Learning Conditional Random Fields for Activity Recognition – GPS to street map association – Inferring activities and types of significant places – Place detection and labeling algorithm Experimental Results – Experimental environment – Example analysis – Extracting significant places – Labeling places and activities using models learned form others Conclusions

Motivation (cont’) Application areas of learning patterns of human behavior from sensor data – Intelligent environments – Surveillance – Human robot interaction Using GPS location data to learn to recognize the high-level activities Difficulties in previous approaches – Restricted activity models – Inaccurate place detection

Motivation A novel, unified approach to automated activity and place labeling – High accuracy in detecting significant places by taking a user’s context into account – By simultaneously using CRF (Conditional Random Field) Estimating a person’s activities Identifying places Labeling places by their type Research goal – To segment a user’s day into everyday activities – To recognize and label significant places

Hierarchical activity model (cont’) GPS readings – Input to proposing model – Segmenting a GPS trace spatially in order to generate a discrete sequence of activity nodes Activities – Being estimated for each node in the spatially segmented GPS trace – Distinguishing between navigation activities and significant activities Significant places – Playing a significant role in the activities of a person

Hierarchical activity model Two key problems for probabilistic inference – Complexity of model Solved by approximating inference algorithm – Not clear how to construct the model deterministically from a GPS trace Solved by constructing the model as part of this inference

Preliminaries : Conditional Random Fields

Overview (cont’) Definition of CRFs – Undirected graphical models developed for labeling sequence data – Properties Directly represent the conditional distribution over hidden states No assumptions about the dependency structure between observations Nodes in CRFs – Observation : – Hidden states : – Defining conditional distribution over hidden states y Cliques – Fully connected sub-graphs of a CRF – Playing a key role in the definition of conditional distribution Preliminaries: Conditional random fields

Overview Conditional distribution over hidden state : where Preliminaries: Conditional random fields

Inference (cont’) Inference in CRF can have two tasks : – To estimate the marginal distribution of each hidden variable – To estimate the most likely configuration of the hidden variables (i.e. the maximum a posteriori, or MAP, estimation) – Using Belief propagation to solve these tasks Two types of BP algorithms : – Sum-product for marginal estimation – Max-product for MAP estimation Preliminaries: Conditional random fields

Inference (cont’) Sum-product for marginal estimation – Message initialization : Initializing all messages as uniform distr. over – Message update rule : – Message update order : Iterating the message update rule until it (possibly) converges – Convergence conditions : – After convergence, calculation of marginals Preliminaries: Conditional random fields

Inference Max-product for MAP estimation – Very similar to the sum-product – Replaced summation with maximization in the message update rule – After convergence, calculating the MAP belief – Then, each component of Preliminaries: Conditional random fields

Parameter learning (cont’) Goal of parameter learning – To determine the weights of the feature functions – Learn the weights discriminatively Two method – Maximum likelihood (ML) estimation – Maximum pseudo-likelihood (MPL) estimation Parameter sharing – Learning algorithm to learn the same parameter values (weights) for different cliques in the CRF Preliminaries: Conditional random fields

Parameter learning (cont’) Maximum likelihood (ML) estimation – Object function – The gradient of object function Preliminaries: Conditional random fields

Parameter learning (cont’) Maximum pseudo-likelihood (MPL) estimation : local feature counts involving variable – Object function – The gradient of object function Preliminaries: Conditional random fields

Parameter learning Parameter sharing – Learn a generic model that can take any GPS trace and classify the locations in that trace – Achieved by making sure that all the weights belonging to a certain type of feature are identical – Calculating gradient for a shared weight by the sum of all the gradients computed for the individual cliques Preliminaries: Conditional random fields

Conditional Random Fields for Activity Recognition

GPS to street map association (cont’) Desirable to associate GPS traces to a street map – (e.g.) to relate locations to addresses in the map Constructing a CRF – Taking into account the spatial relationship between GPS readings – Generating a consistent association Conditional Random Fields for Activity Recognition

GPS to street map association (cont’) Distinguishing tree types of cliques – Measurement cliques (dark grey) – Consistency cliques (light grey) – Smoothness cliques (medium grey) Conditional Random Fields for Activity Recognition

GPS to street map association Conditional Random Fields for Activity Recognition

Inferring activities and types of significant places (cont’) Generating a new CRF, to estimate – Activity performed at each segment – A person’s significant places Conditional Random Fields for Activity Recognition

Inferring activities and types of significant places Activity node’s features – Temporal information such as time of day, day of week, duration of the stay – Average speed through a segment – Information extracted from geographic databases – Connected to its neighbors Place node’s feature – Activities that occur at a place strongly (consider weekly frequency) – A limited number of different homes or work places Possibility of generating very large cliques – Resolve this problem by converting to tree-structured CRFs Conditional Random Fields for Activity Recognition

Place detection and labeling algorithm Conditional Random Fields for Activity Recognition

Experimental Results

Experimental environment Collected GPS data from four different persons – Seven days of data – Roughly 40,000 GPS measurements (10,000 segments) – Manually labeled all activities and significant places Using leave-one-out cross-validation for evaluation – Training data : 3 persons (MPL estimation for learning) – Testing data : 4 persons Experimental Results

Example analysis Experimental Results

Extracting significant places Comparing experiment – Proposing system – A widely-used approach (time threshold) Experimental Results

Labeling places and activities using models learned form others (cont’) Experimental Results

Labeling places and activities using models learned form others Experimental Results

Conclusions A novel approach to performing location-based activity recognition – One consistent framework – Iteratively constructing a hierarchical CRF – Discriminative learning using pseudo-likelihood – Being performed the Inference efficiently using loopy BP Achieving virtually identical accuracy both with and without a street map