Understanding Human Behavior from Sensor Data

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

Opportunity Knocks: A Community Navigation Aid Henry Kautz Don Patterson Dieter Fox Lin Liao University of Washington Computer Science & Engineering.
Recognizing Human Activity from Sensor Data Henry Kautz University of Washington Computer Science & Engineering graduate students: Don Patterson, Lin Liao.
Dynamic Bayesian Networks (DBNs)
Assisted Cognition Henry Kautz University of Rochester Computer Science.
The Future of Computing CSC 161: The Art of Programming Prof. Henry Kautz 12/2/
1 Assisted Cognition Henry Kautz, Oren Etzioni, & Dieter Fox University of Washington Department of Computer Science & Engineering.
Dieter Fox University of Washington Department of Computer Science & Engineering.
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Hardness-Aware Restart Policies Yongshao Ruan, Eric Horvitz, & Henry Kautz IJCAI 2003 Workshop on Stochastic Search.
Personalized Medicine Research at the University of Rochester Henry Kautz Department of Computer Science.
Bayesian Filtering for Robot Localization
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Chapter 8 Prediction Algorithms for Smart Environments
Robust Activity Recognition Henry Kautz University of Washington Computer Science & Engineering graduate students: Don Patterson, Lin Liao, Krzysztof Gajos,
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
Mapping and Localization with RFID Technology Matthai Philipose, Kenneth P Fishkin, Dieter Fox, Dirk Hahnel, Wolfram Burgard Presenter: Aniket Shah.
Inferring High-Level Behavior from Low-Level Sensors Don Peterson, Lin Liao, Dieter Fox, Henry Kautz Published in UBICOMP 2003 ICS 280.
Recognizing Activities of Daily Living from Sensor Data Henry Kautz Department of Computer Science University of Rochester.
DEPARTMENT of COMPUTER SCIENCE University of Rochester  Activities  Abductive Inference of Multi-Agent Interaction  Capture the Flag Data Collection.
Learning and Inferring Transportation Routines By: Lin Liao, Dieter Fox and Henry Kautz Best Paper award AAAI’04.
Fast Simulators for Assessment and Propagation of Model Uncertainty* Jim Berger, M.J. Bayarri, German Molina June 20, 2001 SAMO 2001, Madrid *Project of.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Advisor: Hsin-His Chen Reporter: Chi-Hsin Yu Date: From AAAI 2008 William Pentney, Department of Computer Science & Engineering University of.
Learning to Navigate Through Crowded Environments Peter Henry 1, Christian Vollmer 2, Brian Ferris 1, Dieter Fox 1 Tuesday, May 4, University of.
CS Statistical Machine learning Lecture 24
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
The New Age of Commonsense Reasoning Henry Kautz University of Rochester.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Inferring High-Level Behavior from Low-Level Sensors Donald J. Patterson, Lin Liao, Dieter Fox, and Henry Kautz.
Balance and Filtering in Structured Satisfiability Problems Henry Kautz University of Washington joint work with Yongshao Ruan (UW), Dimitris Achlioptas.
Project ACCESS Henry Kautz, Dieter Fox, Gaetano Boriello (UW CSE) Don Patterson (UW / UC Irvine) Kurt Johnson, Pat Brown, Mark Harniss (UW Rehabilitation.
Sensor Analysis – Part II A literature based exploration Thomas Plötz [material taken from the original papers and John Krumm “Ubiquitous Computing Fundamentals”
AAAI of 20 Deconstructing Planning as Satisfiability Henry Kautz University of Rochester in collaboration with Bart Selman and Jöerg Hoffmann.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Thrust IIA: Environmental State Estimation and Mapping Dieter Fox (Lead) Nicholas Roy MURI 8 Kickoff Meeting 2007.
Learning and Inferring Transportation Routines Lin Liao, Don Patterson, Dieter Fox, Henry Kautz Department of Computer Science and Engineering University.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Dan Roth University of Illinois, Urbana-Champaign 7 Sequential Models Tutorial on Machine Learning in Natural.
Chapter 6 Activity Recognition from Trajectory Data Yin Zhu, Vincent Zheng and Qiang Yang HKUST November 2011.
Assisted Cognition Systems Henry Kautz Department of Computer Science.
Keep the Adversary Guessing: Agent Security by Policy Randomization
Intelligent Transportation System
What is a Hidden Markov Model?
Machine Learning for Computer Security
Understanding Human Behavior from Sensor Data
Understanding Human Behavior from Sensor Data
Constrained Hidden Markov Models for Population-based Haplotyping
Ubiquitous Computing and Augmented Realities
Tracking Objects with Dynamics
Probabilistic Robotics
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Vijay Srinivasan, John Stankovic, Kamin Whitehouse
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18
Probability and Time: Markov Models
Probability and Time: Markov Models
Probabilistic Models with Latent Variables
Probability and Time: Markov Models
Discovering Activities of Daily Life Using RFID’s
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Expectation-Maximization & Belief Propagation
Probability and Time: Markov Models
Finding Periodic Discrete Events in Noisy Streams
Speech recognition, machine learning
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Speech recognition, machine learning
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Understanding Human Behavior from Sensor Data Henry Kautz University of Washington

A Dream of AI Systems that can understand ordinary human experience Work in KR, NLP, vision, IUI, planning… Plan recognition Behavior recognition Activity tracking Goals Intelligent user interfaces Step toward true AI

Plan Recognition, circa 1985

Behavior Recognition, circa 2005

Punch Line Resurgence of work in behavior understanding, fueled by Advances in probabilistic inference Graphical models Scalable inference KR U Bayes Ubiquitous sensing devices RFID, GPS, motes, … Ground recognition in sensor data Healthcare applications Aging boomers – fastest growing demographic

This Talk Activity tracking from RFID tag data ADL Monitoring Learning patterns of transportation use from GPS data Activity Compass Learning to label activities and places Life Capture Other recent research

This Talk Activity tracking from RFID tag data ADL Monitoring Learning patterns of transportation use from GPS data Activity Compass Learning to label activities and places Life Capture Other recent research

Object-Based Activity Recognition Activities of daily living involve the manipulation of many physical objects Kitchen: stove, pans, dishes, … Bathroom: toothbrush, shampoo, towel, … Bedroom: linen, dresser, clock, clothing, … We can recognize activities from a time-sequence of object touches

Image Courtesy Intel Research Application ADL (Activity of Daily Living) monitoring for the disabled and/or elderly Changes in routine often precursor to illness, accidents Human monitoring intrusive & inaccurate Image Courtesy Intel Research

Sensing Object Manipulation RFID: Radio-frequency identification tags Small No batteries Durable Cheap Easy to tag objects Near future: use products’ own tags

Wearable RFID Readers 13.56MHz reader, radio, power supply, antenna 12 inch range, 12-150 hour lifetime Objects tagged on grasping surfaces

Experiment: ADL Form Filling Tagged real home with 108 tags 14 subjects each performed 12 of 65 activities (14 ADLs) in arbitrary order Used glove-based reader Given trace, recreate activities

Results: Detecting ADLs Activity Prior Work SHARP Personal Appearance 92/92 Oral Hygiene 70/78 Toileting 73/73 Washing up 100/33 Appliance Use 100/75 Use of Heating 84/78 Care of clothes and linen 100/73 Making a snack 100/78 Making a drink 75/60 Use of phone 64/64 Leisure Activity 100/79 Infant Care 100/58 Medication Taking 100/93 Housework 100/82 Legend Point solution General solution Inferring ADLs from Interactions with Objects Philipose, Fishkin, Perkowitz, Patterson, Hähnel, Fox, and Kautz IEEE Pervasive Computing, 4(3), 2004

Experiment: Morning Activities 10 days of data from the morning routine in an experimenter’s home 61 tagged objects 11 activities Often interleaved and interrupted Many shared objects Use bathroom Make coffee Set table Make oatmeal Make tea Eat breakfast Make eggs Use telephone Clear table Prepare OJ Take out trash

Baseline: Individual Hidden Markov Models 68% accuracy 11.8 errors per episode

Baseline: Single Hidden Markov Model 83% accuracy 9.4 errors per episode

Cause of Errors Observations were types of objects Spoon, plate, fork … Typical errors: confusion between activities Using one object repeatedly Using different objects of same type Critical distinction in many ADL’s Eating versus setting table Dressing versus putting away laundry

Aggregate Features HMM with individual object observations fails No generalization! Solution: add aggregation features Number of objects of each type used Requires history of current activity performance DBN encoding avoids explosion of HMM

Dynamic Bayes Net with Aggregate Features 88% accuracy 6.5 errors per episode

Improving Robustness DBN fails if novel objects are used Solution: smooth parameters over abstraction hierarchy of object types

Abstraction Smoothing Methodology: Train on 10 days data Test where one activity substitutes one object Change in error rate: Without smoothing: 26% increase With smoothing: 1% increase

Summary Activities of daily living can be robustly tracked using RFID data Simple, direct sensors can often replace (or augment) general machine vision Accurate probabilistic inference requires sequencing, aggregation, and abstraction Works for essentially all ADLs defined in healthcare literature D. Patterson, H. Kautz, & D. Fox, ISWC 2005 best paper award

This Talk Activity tracking from RFID tag data ADL Monitoring Learning patterns of transportation use from GPS data Activity Compass Learning to label activities and places Life Capture Other recent research

Motivation: Community Access for the Cognitively Disabled

The Problem Using public transit cognitively challenging Learning bus routes and numbers Transfers Recovering from mistakes Point to point shuttle service impractical Slow Expensive Current GPS units hard to use Require extensive user input Error-prone near buildings, inside buses No help with transfers, timing

Solution: Activity Compass User carries smart cell phone System infers transportation mode GPS – position, velocity Mass transit information Over time system learns user model Important places Common transportation plans Mismatches = possible mistakes System provides proactive help

Technical Challenge Given a data stream from a GPS unit... Infer the user’s mode of transportation, and places where the mode changes Foot, car, bus, bike, … Bus stops, parking lots, enter buildings, … Learn the user’s daily pattern of movement Predict the user’s future actions Detect user errors

Technical Approach Map is a directed graph G=(V,E) Location: Movement: Edge e Distance d from start of edge Actual (displaced) GPS reading Movement: Mode { foot, car, bus } determines velocity range Change mode near bus stops & parking places Tracking (filtering): Given some prior estimate, Update position & mode according to motion model Correct according to next GPS reading So far, I have described the general picture of our system and some applications. In the rest of the presentation, I will focus on the technical approach of probabilistic reasoning. There are three key things in our system. For the .., we use…. Because of the time issue, I cannot discuss those techniques in details. So I will only give you a very rough picture for each of them and show the results of our experiments.

Dynamic Bayesian Network I mk-1 mk Transportation mode xk-1 xk Edge, velocity, position qk-1 qk Data (edge) association zk-1 zk GPS reading Time k-1 Time k

Mode & Location Tracking Measurements Projections Bus mode Car mode Foot mode Green Red Blue Explain why it switches from foot to bus, not car: because of the bus stops If people ask why no car particle at all (there are car particles, just their weights are too small that can not be displayed).

Learning Prior knowledge – general constraints on transportation use Vehicle speed range Bus stops Learning – specialize model to particular user 30 days GPS readings Unlabeled data Learn edge transition parameters using expectation-maximization (EM)

How to improve predictive power? Predictive Accuracy How to improve predictive power? Probability of correctly predicting the future City Blocks

Transportation Routines Home A B Workplace Goal: intended destination Workplace, home, friends, restaurants, … Trip segments: <start, end, mode> Home to Bus stop A on Foot Bus stop A to Bus stop B on Bus Bus stop B to workplace on Foot Now the system could infer location and transportation modes, but wee still want to know more. We first show an example. Goal In order to get to the goal, the user may switch transportation mode. Thus we introduce the concept of trip segments.

Dynamic Bayesian Net II gk-1 gk Goal tk-1 tk Trip segment mk-1 mk Transportation mode xk-1 xk Edge, velocity, position qk-1 qk Data (edge) association zk-1 zk GPS reading Time k-1 Time k

Unsupervised Hierarchical Learning Use previous model to infer: Goals locations where user stays for a long time Transition points locations with high mode transition probability Trip segments paths connecting transition points or goals Learn transition probabilities Lower levels conditioned on higher levels Expectation-Maximization

Predict Goal and Path Predicted goal Predicted path

Improvement in Predictive Accuracy

Detecting User Errors Learned model represents typical correct behavior Model is a poor fit to user errors We can use this fact to detect errors! Cognitive Mode Normal: model functions as before Error: switch in prior (untrained) parameters for mode and edge transition After we learn the model, detect potential errors.

Dynamic Bayesian Net III zk-1 zk xk-1 xk qk-1 qk Time k-1 Time k mk-1 mk tk-1 tk gk-1 gk ck-1 ck Cognitive mode { normal, error } Goal Trip segment Transportation mode Edge, velocity, position Data (edge) association GPS reading

Detecting User Errors In this episode, the person takes the bus home but missed the usual bus stop. The black disk is the goal and the cross mark is where the system think the person should get off the bus. The probability of normal and abnormal activities are represented using a bar here. The black part is the probability of being normal while the red part is the probability of being abnormal. At here, the person missed the bus stop and the probability of abnormal activity jump quickly.

Status Major funding by NIDRR Extension to indoor navigation National Institute of Disability & Rehabilitation Research Partnership with UW Dept of Rehabilitation Medicine & Center for Technology and Disability Studies Extension to indoor navigation Hospitals, nursing homes, assisted care communities Wi-Fi localization Multi-modal interface studies Speech, graphics, text Adaptive guidance strategies

Papers Patterson, Liao, Fox, & Kautz, UBICOMP 2003 Inferring High Level Behavior from Low Level Sensors Patterson et al, UBICOMP-2004 Opportunity Knocks: a System to Provide Cognitive Assistance with Transportation Services Liao, Fox, & Kautz, AAAI 2004 (Best Paper) Learning and Inferring Transportation Routines

This Talk Activity tracking from RFID tag data ADL Monitoring Learning patterns of transportation use from GPS data Activity Compass Learning to label activities and places Life Capture Other recent research

Task Learn to label a person’s Given Daily activities working, visiting friends, traveling, … Significant places work place, friend’s house, usual bus stop, … Given Training set of labeled examples Wearable sensor data stream GPS, acceleration, ambient noise level, …

Applications Activity Compass Life Capture Automatically assign names to places Life Capture Automated diary

Conditional Models HMMs and DBNs are generative models Describe complete joint probability space For labeling tasks, conditional models are often simpler and more accurate Learn only P( label | observations ) Fewer parameters than corresponding generative model

Things to be Modeled Raw GPS reading (observed) Actual user location Activities (time dependent) Significant places (time independent) Soft constraints between all of the above (learned)

Conditional Random Field Undirected graphical model Feature functions defined on cliques Conditional probability proportional to exp( weighted sum of features ) Weights learned by maximizing (pseudo) likelihood of training data

Relational Markov Network First-order version of conditional random field Features defined by feature templates All instances of a template have same weight Examples: Time of day an activity occurs Place an activity occurs Number of places labeled “Home” Distance between adjacent user locations Distance between GPS reading & nearest street

Global soft constraints RMN Model s1 Global soft constraints p1 p2 Significant places a1 a2 a3 a4 a5 Activities g1 g2 g3 g4 g5 g6 g7 g8 g9 {GPS, location} time 

Significant Places Previous work decoupled identifying significant places from rest of inference Simple temporal threshold Misses places with brief activities RMN model integrates Identifying significant place Labeling significant places Labeling activities

Efficient Inference Some features are expensive to handle by general inference algorithms E.g. belief propagation, MCMC Can dramatically speed up inference by associating inference procedures with feature templates Fast Fourier transform (FFT) to compute “counting” features O(n log2n) versus O(2n)

Results: Labeling One user, 1 week training data, 1 week testing data Number of (new) significant places correctly labeled: 18 out of 19 Number of activities correctly labeled: 53 out of 61 Number of activities correctly labeled, if counting features not used: 44 out of 61

Results: Finding Significant Places

Results: Efficiency

Summary We can learn to label a user’s activities and meaningful locations using sensor data & state of the art relational statistical models (Liao, Fox, & Kautz, IJCAI 2005, NIPS 2006) Many avenues to explore: Transfer learning Finer grained activities Structured activities Social groups

This Talk Activity tracking from RFID tag data ADL Monitoring Learning patterns of transportation use from GPS data Activity Compass Learning to label activities and places Life Capture Other recent research

Planning as Satisfiability Unless P=NP, no polynomial time algorithm for SAT But: great practical progress in recent years 1980: 100 variable problems 2005: 100,000 variable problems Can we use SAT as a engine for planning? 1996 – competitive with state of the art ICAPS 2004 Planning Competition – 1st prize, optimal STRIPS planning Inspired research on bounded model-checking

Understanding Clause Learning Modern SAT solvers extend Davis-Putnam backtrack search with clause learning Cache “reason” for each backtrack as an inferred clause Works in practice, but not clear why… Understanding clause learning using proof complexity (Sabharwal, Beame, & Kautz 2003, 2004, 2005) More powerful than tree-like resolution Separation from regular resolution Increase practical benefit by domain-specific variable- selection heuristics Linear-size proofs of pebbling graphs

Efficient Model Counting SAT – can a formula be satisfied? #SAT – how many ways can a formula be satisfied? Compact translation discrete Bayesian networks  #SAT Efficient model counting (Sang, Beame, & Kautz 2004, 2005) Formula caching Component analysis New branching heuristics Cachet – fastest model counting algorithm

Learning Control Policies for Satisfiability Solvers High variance in runtime distributions of SAT solvers Over different random seeds for tie-breaking in branching heuristics Sometimes: infinite variance (heavy-tailed) Restarts can yield super-linear speedup Can learn optimal restart policies using features of instance + solver trace (Ruan, Kautz, & Horvitz 2001, 2002, 2003)

Other Projects Continuous time Bayesian networks (Gopalratnam, Weld, & Kautz, AAAI 2005) Building social network models from sensor data Probabilistic speech act theory Modal Markov Logic

Conclusion: Why Now? An early goal of AI was to create programs that could understand ordinary human experience This goal proved elusive Missing probabilistic tools Systems not grounded in real world Lacked compelling purpose Today we have the mathematical tools, the sensors, and the motivation

Credits Graduate students: Colleagues: Funders: Don Patterson, Lin Liao, Ashish Sabharwal, Yongshao Ruan, Tian Sang, Harlan Hile, Alan Liu, Bill Pentney, Brian Ferris Colleagues: Dieter Fox, Gaetano Borriello, Dan Weld, Matthai Philipose Funders: NIDRR, Intel, NSF, DARPA