Learning to Satisfy Actuator Networks

Slides:



Advertisements
Similar presentations
Bayesian Belief Propagation
Advertisements

CS188: Computational Models of Human Behavior
Autonomic Scaling of Cloud Computing Resources
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for 1 Lecture Notes for E Alpaydın 2010.
Factorial Mixture of Gaussians and the Marginal Independence Model Ricardo Silva Joint work-in-progress with Zoubin Ghahramani.
Adopt Algorithm for Distributed Constraint Optimization
Supervised Learning Recap
Decision Making: An Introduction 1. 2 Decision Making Decision Making is a process of choosing among two or more alternative courses of action for the.
Hidden Markov Models M. Vijay Venkatesh. Outline Introduction Graphical Model Parameterization Inference Summary.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Visual Recognition Tutorial
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
x – independent variable (input)
Hidden Markov Model 11/28/07. Bayes Rule The posterior distribution Select k with the largest posterior distribution. Minimizes the average misclassification.
Bayesian Networks I: Static Models & Multinomial Distributions By Peter Woolf University of Michigan Michigan Chemical Process Dynamics.
Energy Management and Adaptive Behavior Tarek Abdelzaher.
Dealing with NP-Complete Problems
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Visual Recognition Tutorial
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Crash Course on Machine Learning
Bayes Net Perspectives on Causation and Causal Inference
Incomplete Graphical Models Nan Hu. Outline Motivation K-means clustering Coordinate Descending algorithm Density estimation EM on unconditional mixture.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Machine Learning and Optimization For Traffic and Emergency Resource Management. Milos Hauskrecht Department of Computer Science University of Pittsburgh.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
Sergios Theodoridis Konstantinos Koutroumbas Version 2
Daniel Guetta (DRO)Transitional Care Units IEOR Final Project 9 th May 2012 Daniel Guetta Joint work with Carri Chan.
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
Lecture 2: Statistical learning primer for biologists
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Robust Estimation With Sampling and Approximate Pre-Aggregation Author: Christopher Jermaine Presented by: Bill Eberle.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
Distributed cooperation and coordination using the Max-Sum algorithm
Optimization-based Cross-Layer Design in Networked Control Systems Jia Bai, Emeka P. Eyisi Yuan Xue and Xenofon D. Koutsoukos.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
Nevin L. Zhang Room 3504, phone: ,
12. Principles of Parameter Estimation
Empirical risk minimization
LECTURE 03: DECISION SURFACES
Financial Econometrics Lecture Notes 4
Oliver Schulte Machine Learning 726
Bounding the Lifetime of Sensor Networks Via Optimal Role Assignments
Precision Agriculture an Overview
Risk-informed Decision Making under Incomplete Information
Towards Next Generation Panel at SAINT 2002
Hidden Markov Models Part 2: Algorithms
More about Posterior Distributions
Bayesian Models in Machine Learning
CSCI 5822 Probabilistic Models of Human and Machine Learning
Bayesian Networks Independencies Representation Probabilistic
An Algorithm for Bayesian Network Construction from Data
Nonparametric Hypothesis Tests for Dependency Structures
Empirical risk minimization
Biointelligence Laboratory, Seoul National University
12. Principles of Parameter Estimation
Kostas Kolomvatsos, Christos Anagnostopoulos
Presentation transcript:

Learning to Satisfy Actuator Networks Mark Coates National Science and Engineering Research Council of Canada (NSERC)

A Journey “And what is a journey?  Is it just… distance traveled?  Time spent?  No.  It's what happens on the way, the things that happen to you.  At the end of the journey you're not the same.”

Plan your Journey, Learn “It's what happens on the way, it the things that happen to you.” Sensor Networks SANETs Local Actuation: Control sensors, control objects. Modify the environment. Learn causal relationships between actuations and environmental (model) variables. Plan behaviour to optimize performance

Sensor/Actuator Network Set of actuators ( ) + associated sensors ( ). Actuators perform a physical (modifying) action. Sensors monitor the response of the system. Quantify the net effect on the system (positive or negative) Design actuation strategy to optimize response

Causal Analysis How can we infer the impact of an actuation based on a set of observations? In particular, how do we derive: Manipulated Probability P(Y | X := x, Z=z) From (observations based on) Unmanipulated Probability P(Y | X = x, Z=z)

Example Problem We wish to evaluate the average effectiveness of a fertilizer Local background variables (for example soil moisture, temperature, salinity, weed density) affect: The successful reception of the fertilizer The impact of the fertilizer on the crop

Causal Graph uj : local realizations of background variables with global distribution g zj : action by actuator (0/1 = off/on) [known or measured] dj : actuation received (0/1 = no/yes) [unobserved] yj : response (0/1 = negative/positive) [unobserved] xj , wj: observed measurements, dependent on dj and yj.

Average Causal Effect (ACE) Expectation (over latent variables) of: [ Prob. of positive response given fertilizer ― Prob. of positive response without fertilizer ]

Model the mapping not the variable Problem: Latent variables u can be high dimensional; Probability distribution g(u) can have complex structure. Approach: We have binary variables Z, D, Y We don’t care about the value of uj and how that directly influences dj What we do care about is how u impacts the mappings ZD and DY

New Causal Graph cr: sixteen states c= Much easier to estimate this distribution 0: inhibit 1: pass 2: flip 3: activate crj zj dj xj yj wj

Evaluating ACE Estimate ACE by applying distributed EM algorithm across the graphical model (model g(cr) as multinomial) Locally maximize the likelihood function: Expectation: calculate expected crj at each node Maximization: average the expected cri to estimate g(cr).

Sensor Network Evaluation Tree network topology: An efficient mechanism for data aggregation and dissemination. Data aggregation (bottom-up) Leaf nodes: Transmit E[crj] to parent node Parent node: Performs aggregation and relays result to its own parent Root node: Performs maximization Result dissemination (top-down) Each node broadcasts result to its children nodes

Influencing the Environment Design an actuation strategy Set of decision rules Map from (current and past) sensor measurements to an actuation Possible Objectives: Maximize expected response of system Provide probabilistic bounds on worst-case behaviour Possible Scenarios Accurate models of probability distributions  Bayesian networks Uninformative models  Learning approaches

Problem Formulation Epoch of T discrete time intervals At times t = 0,…,T, node i measures a set of environmental variables Chooses an actuation belonging to a discrete set of actuations At the end of the epoch, measure a local response variable

Maximization Approach Single binary actuation decision without a good model of p(y|v,a) Consider p(y|v,a) = f(y|v,a) + n(v,a) We have a set of points (vi,ai,yi) and want to learn the best actuation strategy A(v), i.e., that which maximizes f(y|v,a). Approach: regression + subsequent maximization

Robustness concerns Maximization amplifies regression errors. Multi-stage planning implies repeated regression + maximization Proliferation of error

Relaxed Problem Identify the largest set of environmental conditions and actuations such that: Expected response exceeds threshold Probability of terrible response is very low

LSAT Formulation Learning to Satisfy (LSAT) Given points find the set G that solves: subject to where Ci(G,P) are constraints.

Two types of constraints Point-wise constraints: C(G,P) = C(x,G,P) is a function of the input variable x and the constraint takes the form Example: Set-average constraints: C(G,P) > 0 is only satisfied on-average over the entire set.

Solution Derive equivalent empirical constraints Consider solution to empirical constraints subject to If empirical constraints are close to ideal constraints then solution is satisfactory. Algorithm: extension of support vector machine Lagrangian formulation allocating different penalties to violations of individual constraints.

Comments Actuator networks present a host of problems Assessment of whether causal relationships exist and evaluation of their strength Design of actuation strategies that yield a satisfactory (optimal?) environmental response These problems are difficult in a centralized setting – the extension to distributed algorithms poses an even greater challenge.