Exploiting Graphical Structure in Decision-Making

Slides:



Advertisements
Similar presentations
Learning Trajectory Patterns by Clustering: Comparative Evaluation Group D.
Advertisements

Continuation Methods for Structured Games Ben Blum Christian Shelton Daphne Koller Stanford University.
Computer vision: models, learning and inference Chapter 8 Regression.
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
Temporal Action-Graph Games: A New Representation for Dynamic Games Albert Xin Jiang University of British Columbia Kevin Leyton-Brown University of British.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
Max-norm Projections for Factored MDPs Carlos Guestrin Daphne Koller Stanford University Ronald Parr Duke University.
CPSC 322, Lecture 12Slide 1 CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12 (Textbook Chpt ) January, 29, 2010.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Multiagent Planning with Factored MDPs Carlos Guestrin Daphne Koller Stanford University Ronald Parr Duke University.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
ADITI BHAUMICK ab3585. To use reinforcement learning algorithm with function approximation. Feature-based state representations using a broad characterization.
A1A1 A4A4 A2A2 A3A3 Context-Specific Multiagent Coordination and Planning with Factored MDPs Carlos Guestrin Shobha Venkataraman Daphne Koller Stanford.
MAKING COMPLEX DEClSlONS
Research, Development, Consulting, Training High Fidelity Modeling and Simulation Where we are going… …future plans.
Representing and Using Graphs
CSC401: Analysis of Algorithms CSC401 – Analysis of Algorithms Chapter Dynamic Programming Objectives: Present the Dynamic Programming paradigm.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 11 th, 2006 Readings: K&F: 8.1, 8.2, 8.3,
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
MDPs (cont) & Reinforcement Learning
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Static Process Scheduling
Instructor: Spyros Reveliotis IE7201: Production & Service Systems Engineering Fall 2009 Closure.
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Introduction and Preliminaries D Nagesh Kumar, IISc Water Resources Planning and Management: M4L1 Dynamic Programming and Applications.
1 Variable Elimination Graphical Models – Carlos Guestrin Carnegie Mellon University October 15 th, 2008 Readings: K&F: 8.1, 8.2, 8.3,
A Brief Introduction to Bayesian networks
Lecture 7: Constrained Conditional Models
Uniformed Search (cont.) Computer Science cpsc322, Lecture 6
Recap of L09: Normative Decision Theory
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Maximum Expected Utility
Author: Vikas Sindhwani and Amol Ghoting Presenter: Jinze Li
New Characterizations in Turnstile Streams with Applications
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
ECE 448 Lecture 4: Search Intro
Uniformed Search (cont.) Computer Science cpsc322, Lecture 6
NP-Completeness Yin Tat Lee
Encoding CNFs to Enhance Component Analysis
CSPs: Search and Arc Consistency Computer Science cpsc322, Lecture 12
Structured Models for Multi-Agent Interactions
Turnstile Streaming Algorithms Might as Well Be Linear Sketches
Dynamic Programming General Idea
CMSC 471 Fall 2009 RL using Dynamic Programming
Chapter 4: Dynamic Programming
Estimating Networks With Jumps
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
Goodfellow: Chapter 14 Autoencoders
Chapter 4: Dynamic Programming
Readings: K&F: 15.1, 15.2, 15.3, 15.4, 15.5 K&F: 7 (overview of inference) K&F: 8.1, 8.2 (Variable Elimination) Structure Learning in BNs 3: (the good,
Greedy Algorithms Many optimization problems can be solved more quickly using a greedy approach The basic principle is that local optimal decisions may.
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CIS 488/588 Bruce R. Maxim UM-Dearborn
Raphael Yuster Haifa University Uri Zwick Tel Aviv University
Dynamic Programming Dynamic Programming 1/18/ :45 AM
Dynamic Programming Merge Sort 1/18/ :45 AM Spring 2007
General Gibbs Distribution
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11
Chapter 8: Generalization and Function Approximation
NP-Completeness Yin Tat Lee
Dynamic Programming General Idea
Lecture 6 Dynamic Programming
Merge Sort 4/28/ :13 AM Dynamic Programming Dynamic Programming.
CS 416 Artificial Intelligence
Chapter 4: Dynamic Programming
Variable Elimination Graphical Models – Carlos Guestrin
Dynamic Programming Merge Sort 5/23/2019 6:18 PM Spring 2008
A task of induction to find patterns
Reinforcement Learning (2)
Reinforcement Learning (2)
Presentation transcript:

Exploiting Graphical Structure in Decision-Making Ben Van Roy Stanford University

Overview Graphical models in decision-making Singly-connected  efficient computation General decision problems  intractable Sparsity  reduction in computation? Sequential decision-making Curse of dimensionality Sparsity or other graphical structure  reduced computational requirements? Structured value functions and/or policies? Preliminary results and research directions

Graphical Models in Inference Conditional independencies simplify inference Singly-connected graphs (trees) General sparse graphs Preprocessing Approximations? x1 x2 x3 x4

Graphical Models in Decision-Making Deterministic dynamic programming Nonserial dynamic programming General sparse graphs Preprocessing Approximations? x1 x2 x3 x4

Sequential Decision-Making u(t) state x(t) system strategy Bellman’s equation J*(x(t)) = max E[g(x(t), u(t)) + a J*(x(t+1)) | x(t)]

The Curse of Dimensionality # states is exponential in # variables The value function encodes one value per state Storage is intractable Computation is intractable Research objective: exploit sparsity and other special graphical structure to reduce computational requirements of sequential decision problems

Dynamic Bayesian Networks x1(t) x1(t+1) x2(t) x2(t+1) x3(t) x3(t+1) x4(t) x4(t+1)

Example Multiclass Queueing Networks

Can We Exploit Proximity? Idea: variables that are “far” from each don’t interact much Does this allow us to decompose the problem?

Yes… The value function decomposes N(i) = a neighborhood; i.e. a set of nodes within some “distance” of i Complexity: O(nd)  O(dnN) …but there’s a problem here…

Optimal Decisions Depend on Global State information u1(t+1) x1(t) x1(t+1) x2(t) x2(t+1) x3(t) x3(t+1) x4(t) x4(t+1)

Things Still Work Out… Conjecture: Consequence: If decision ui influences only xi Then near-optimal decisions can be made based only on variables “near” xi Consequence: u1(t+1) x1(t) x1(t+1) x2(t) x2(t+1) x3(t) x3(t+1) x4(t) x4(t+1)

The Underlying Problem x7 x2 x3 x1 x6 x4 x5 Which fij’s do I need to know to choose a near-optimal uk (without coordination)?

A Simple Case Let N(i) = nodes within r steps x1 x2 x3 x4 x5 x6 Let N(i) = nodes within r steps Result: loss of optimality ~ O(1/r) Note: amount of information required is independent of the graph size (Rusmevichientong and Van Roy, 2000)

Future Work Extending this result to general graphs Exploring practical implications Expected practical utility: reduction of complexity in approximation algorithms Problem is no longer O(nd) May instead be O(dnr) Still computationally prohibitive, but not exponential in problem size Simplification of decision-supporting information?

More Future Work Current work exploits proximity Many graphs arising in practical problems pose additional special structure (e.g., symmetries, multiple “layers” of relationships, etc.) Can we also exploit such structure? (e.g., are there sometimes appropriate hierarchical representations?)