Towards Model-lite Planning A Proposal For Learning & Planning with Incomplete Domain Models Sungwook Yoon Subbarao Kambhampati Supported by DARPA Integrated.

Slides:



Advertisements
Similar presentations
REVIEW : Planning To make your thinking more concrete, use a real problem to ground your discussion. –Develop a plan for a person who is getting out of.
Advertisements

Computer Science CPSC 322 Lecture 25 Top Down Proof Procedure (Ch 5.2.2)
Big Ideas in Cmput366. Search Blind Search State space representation Iterative deepening Heuristic Search A*, f(n)=g(n)+h(n), admissible heuristics Local.
CLASSICAL PLANNING What is planning ?  Planning is an AI approach to control  It is deliberation about actions  Key ideas  We have a model of the.
State Transition Systems  linear planning  bounded model checking  conditional planning  model checking  state transition description languages: oPDDL.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
“Using Weighted MAX-SAT Engines to Solve MPE” -- by James D. Park Shuo (Olivia) Yang.
Planning Graphs * Based on slides by Alan Fern, Berthe Choueiry and Sungwook Yoon.
TWEAK solving the Sussman anomaly ON(A,B) ON(B,C) ON(C,A) ONTABLE(A) ONTABLE(B) CLEAR(C) CLEAR(B) ARMEMPTY STACK(A,B) STACK(B,C) Two step additions to.
Planning Chapter 10.
Planning Planning is fundamental to “intelligent” behaviour. E.g.
All rights reserved ©L. Manevitz Lecture 61 Artificial Intelligence Planning System L. Manevitz.
Sussman anomaly - analysis The start state is given by: ON(C, A) ONTABLE(A) ONTABLE(B) ARMEMPTY The goal by: ON(A,B) ON(B,C) This immediately leads to.
1  Planning (under complete Knowledge) Intro to Artificial Intelligence.
Planning Russell and Norvig: Chapter 11 Slides adapted from: robotics.stanford.edu/~latombe/cs121/2003/ home.htm.
Artificial Intelligence II S. Russell and P. Norvig Artificial Intelligence: A Modern Approach Chapter 11: Planning.
1 Chapter 4 State-Space Planning. 2 Motivation Nearly all planning procedures are search procedures Different planning procedures have different search.
Planning Some material adapted from slides by Tim Finin,Jean-Claude Latombe, Lise Getoor, and Marie desJardins.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
1 Classical STRIPS Planning Alan Fern * * Based in part on slides by Daniel Weld.
CPSC 322 Introduction to Artificial Intelligence November 19, 2004.
Planning Russell and Norvig: Chapter 11. Planning Agent environment agent ? sensors actuators A1A2A3.
OnT-A onT-B onT-C cl-A cl-C cl-B he Pick-A Pick-B Pick-C onT-A onT-B onT-C cl-A cl-C cl-B he h-A h-B h-C ~cl-A ~cl-B ~cl-C ~he st-A-B st-A-C st-B-A st-B-C.
CPSC 322 Introduction to Artificial Intelligence November 26, 2004.
A: A Unified Brand-name-Free Introduction to Planning Subbarao Kambhampati Jan 28 th My lab was hacked and the systems are being rebuilt.. Homepage is.
CSE 574: Planning & Learning Subbarao Kambhampati 1/17: State Space and Plan-space Planning Office hours: 4:30—5:30pm T/Th.
1 Lecture 12 example (from slides prepared by Prof. J. Rosenchein)
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Planning Where states are transparent and actions have preconditions and effects Notes at
Intro to AI Fall 2002 © L. Joskowicz 1 Introduction to Artificial Intelligence LECTURE 12: Planning Motivation Search, theorem proving, and planning Situation.
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
1 BLACKBOX: A New Approach to the Application of Theorem Proving to Problem Solving Bart Selman Cornell University Joint work with Henry Kautz AT&T Labs.
1 Pertemuan 17 Planning Matakuliah: T0264/Intelijensia Semu Tahun: Juli 2006 Versi: 2/1.
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
1 Planning Chapters 11 and 12 Thanks: Professor Dan Weld, University of Washington.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
Planning Where states are transparent and actions have preconditions and effects Notes at
Planning II CSE 573. © Daniel S. Weld 2 Logistics Reading for Wed Ch 18 thru 18.3 Office Hours No Office Hour Today.
PLANNING Partial order regression planning Temporal representation 1 Deductive planning in Logic Temporal representation 2.
An Introduction to Artificial Intelligence CE Chapter 11 – Planning Ramin Halavati In which we see how an agent can take.
Back Propagation and Representation in PDP Networks Psychology 209 February 6, 2013.
CPS 270: Artificial Intelligence Planning Instructor: Vincent Conitzer.
Synthesizing Robust Plans under Incomplete Domain Models Tuan Nguyen and Subbarao Kambhampati Department of CSE, Arizona State University Minh Do Palo.
Conformant Probabilistic Planning via CSPs ICAPS-2003 Nathanael Hyafil & Fahiem Bacchus University of Toronto.
1 Chapter 16 Planning Methods. 2 Chapter 16 Contents (1) l STRIPS l STRIPS Implementation l Partial Order Planning l The Principle of Least Commitment.
Automated Planning Dr. Héctor Muñoz-Avila. What is Planning? Classical Definition Domain Independent: symbolic descriptions of the problems and the domain.
1/16 Planning Chapter 11- Part1 Author: Vali Derhami.
Robust Planning using Constraint Satisfaction Techniques Daniel Buettner and Berthe Y. Choueiry Constraint Systems Laboratory Department of Computer Science.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Local Search Methods for SAT Geoffrey Levine March 11, 2004.
CS621: Artificial Intelligence Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay Lecture 19: Hidden Markov Models.
Knowledge Repn. & Reasoning Lecture #9: Propositional Logic UIUC CS 498: Section EA Professor: Eyal Amir Fall Semester 2005.
1 Planning Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
An Introduction to Artificial Intelligence CE 40417
Fast Propositional Algorithms for Planning
Planning Chapter 16.
Sequential Stochastic Models
Introduction Contents Sungwook Yoon, Postdoctoral Research Associate
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
An Overview of AI Planning
Planning as Satisfiability with Blackbox
CSE-490DF Robotics Capstone
Planning Chapter
CPSC 322 Introduction to Artificial Intelligence
Probability and Time: Markov Models
CS344 : Introduction to Artificial Intelligence
Russell and Norvig: Chapter 11 CS121 – Winter 2003
Prof. Pushpak Bhattacharyya, IIT Bombay
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Towards Model-lite Planning A Proposal For Learning & Planning with Incomplete Domain Models Sungwook Yoon Subbarao Kambhampati Supported by DARPA Integrated Learning Program

A Planning Problem Towards Model-lite Planning - Sungwook Yoon Suppose you have a super fast planner and a target application. What is the first problem you have to solve? Is it a problem from the application? Domain Engineering is hard  Model-lite Planning

Snapshot of the talk This is a proposal. We formulate learning and planning problems and solution methods for them. We tested our idea on some problems. But the verification is still an undergoing process We propose – Representation for model-lite planning probabilistic logic, incompleteness is quantified Explicit consideration of domain invariant – Learning of the domain model Update of the probability and finding of the new axioms – Planning with the model Deterministic planning domain needs probabilistic planning Most plausible plan that respects the current domain model Towards Model-lite Planning - Sungwook Yoon

Representation Precondition Axiom: p A i, A → pre i Uncertainty is quantified as a probability Effect Axiom: e A i, A → effect i Facilitates learning Towards Model-lite Planning - Sungwook Yoon

Domain Model - Blocksworld 0.9, Pickup (x) -> armempty() 1, Pickup (x) -> clear(x) 1, Pickup (x) -> ontable(x) 0.8, Pickup (x) –> holding(x) 0.8, Pickup (x) -> not armempty() 0.8, Pickup (x) -> not ontable(x) Precondition Axiom: Relates Actions with Current state facts Effect Axiom: Relates Actions with Next state facts Towards Model-lite Planning - Sungwook Yoon

Representation One modeling problem Conjunction of the effect have different semantics, if the probability of each effect is independently specified Add hidden variable, O, (e, A → O), then add deterministic axioms for each effect, (1,O → eff1), (1,O → eff2), … We can alleviate this problem also with explicit domain invariant property Writing explicit domain invariant property is easier than writing initial state generator and a set of operators that respects such property Towards Model-lite Planning - Sungwook Yoon 0.8, Pickup (x) –> holding(x) 0.8, Pickup (x) -> not armempty() 0.8, Pickup (x) -> not ontable(x) Effect Axiom: Relates Actions with Next state facts 1, holding(x) -> not armempty() 1, holding(x) -> not ontable(x) Static Property: Relates Facts in a State

Learning the domain model Given a trajectory of states and actions, S 1,A 1,S 2,A 2, …, S n,A n,S n+1 – We can learn precondition axioms from (S 1,A 1 ), (S 2,A 2 ), …, (S n,A n ) – We can learn effect axioms from (A 1,S 2 ), (A 2,S 3 ), …, (A n,S n+1 ) – We can learn domain invariant properties from each state (S 1 ), …, (S n+1 ) – The weights (probabilities) of the axioms can be updated with simple perceptron update There are readily available package for weighted logic learning – Alchemy (MLN) – Problog Structure learning – Alchemy provides structure learning too – We can also enumerate all the possible axioms (very costly for planning) Towards Model-lite Planning - Sungwook Yoon

Model-lite planning  Probabilistic Planning As stated before, with incomplete domain knowledge, a deterministic planning domain should be treated as a probabilistic domain The resulting plan should be maximally consistent with the current domain model We develop a planning technique for this purpose – A plan that is maximally plausible, given the probabilistic axioms, initial state and goal MPE solution to a Bayes Net problem – Build on plangraph Towards Model-lite Planning - Sungwook Yoon

Probabilistic Plangraph AB A B clear_a clear_b armempty ontable_a ontable_b pickup_a pickup_b clear_a clear_b armempty ontable_a ontable_b holding_a holding_b pickup_a pickup_b stack_a_b stack_b_a clear_a clear_b armempty ontable_a ontable_b holding_a holding_b on_a_b on_b_a noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_holding_a noop_holding_b 0.8 How do we generate a weighted clause? 0.95, pickup_b’ v holding_b Red lines indicate Mutexes 0.8 Domain Invariant Property Can be asserted too Towards Model-lite Planning - Sungwook Yoon

AB A B clear_a clear_b armempty ontable_a ontable_b pickup_a pickup_b clear_a clear_b armempty ontable_a ontable_b holding_a holding_b pickup_a pickup_b stack_a_b stack_b_a clear_a clear_b armempty ontable_a ontable_b holding_a holding_b on_a_b on_b_a noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_holding_a noop_holding_b 0.8 Can we view the probabilistic plangraph as Bayes net? Evidence Variables How we find a solution? MPE (most probabilistic explanation) There are some solvers out there Domain Invariant Property Can be asserted too, 0.9 Towards Model-lite Planning - Sungwook Yoon

MPE as Maxsat There has been a work by James D. Park, AAAI 2002 Set –log(P) as the weight of the clauses A/BP TT0.7 FT0.3 TF0.2 FF0.8 Weighted Clauses -log0.7 -A v –B -log0.3 A V –B -log0.2 –A v B -log0.8 A v B Intuitive explanation Violating the clause is easier for High probability instances Thus the MaxSat Problem Gives you the highest probability instantiations A->B, T T 1, T F 0, InfinityWeight for –A v B, (complies with our intuitive understanding) Towards Model-lite Planning - Sungwook Yoon

AB A B clear_a clear_b armempty ontable_a ontable_b pickup_a pickup_b clear_a clear_b armempty ontable_a ontable_b holding_a holding_b pickup_a pickup_b stack_a_b stack_b_a clear_a clear_b armempty ontable_a ontable_b holding_a holding_b on_a_b on_b_a noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_clear_a noop_clear_b noop_armempty noop_ontable_a noop_ontable_b noop_holding_a noop_holding_b -log0.8 Probabilistic Plangraph to MaxSat Evidence Variables -log0.5 For each probabilistic weight, we give –log(1-p)! That’s it. -log0.8 Domain Invariant Property Can be asserted too, -log0.9 Towards Model-lite Planning - Sungwook Yoon

Exploding Blocksworld Towards Model-lite Planning - Sungwook Yoon

Current Status (ongoing) Learning test – Generated Blocksworld Random Wandering Data and feed them to Alchemy with correct and incorrect axioms – Alchemy found higher weight on the correct axioms and lower weight on the incorrect axioms Planning test – Tested on probabilistic planning problems – Hand tested on a couple of instances of Slippery Gripper Domain Hand encoded the clauses and assigned the weight Put the resulting clauses to MaxSat solve Got desired results – On Exploding Blocksworld Implemented generic MaxSat encoder for probabilistic planning problems Tested on a couple of problems from Exploding Blocksworld Finds desired output frequently (not always) Towards Model-lite Planning - Sungwook Yoon

Summary We can learn precondition axioms and effect axioms separately. – A -> Prec, A->Effect – Facilitates the learning Domain axiom or Invariant Property can be, provided, learned and used explicitly – It is better for domain modeler For planning, we can apply probabilistic plangraph approach – We proposed using MaxSat to solve probabilistic planning problems – Interesting parallel to deterministic planning to SAT Towards Model-lite Planning - Sungwook Yoon

Domain Learning – Related Work Logical Filtering (Chang & Eyal, ICAPS’06) – Update belief state and domain transition model – Experiments involved planning Probabilistic operator learning (Zettlemoyer, Pasula and Kaelbling, AAAI’05) – Experiments involved planning ARMS (Yang, Wu and Jiang, ICAPS ‘05) – No observation besides initial state and goal Towards Model-lite Planning - Sungwook Yoon

Probabilistic Planning in Plangraph – Related Work Pgraphplan, Paragraph Both search plans in the graphplan framework. pGraphplan searches for a consistent plan that maximizes the goal-reaching probability – Forward probability propagation Paragraph searches for a plan that minimizes the cost to reach the goal – Backward plan search Towards Model-lite Planning - Sungwook Yoon