Planning Russell and Norvig: Chapter 11. Planning Agent environment agent ? sensors actuators A1A2A3.

Slides:



Advertisements
Similar presentations
REVIEW : Planning To make your thinking more concrete, use a real problem to ground your discussion. –Develop a plan for a person who is getting out of.
Advertisements

Language for planning problems
CSE391 – 2005 NLP 1 Planning The Planning problem Planning with State-space search.
Planning Module THREE: Planning, Production Systems,Expert Systems, Uncertainty Dr M M Awais.
Planning Module THREE: Planning, Production Systems,Expert Systems, Uncertainty Dr M M Awais.
CLASSICAL PLANNING What is planning ?  Planning is an AI approach to control  It is deliberation about actions  Key ideas  We have a model of the.
Planning Planning is fundamental to “intelligent” behaviour. E.g.
All rights reserved ©L. Manevitz Lecture 61 Artificial Intelligence Planning System L. Manevitz.
Planning Russell and Norvig: Chapter 11 Slides adapted from: robotics.stanford.edu/~latombe/cs121/2003/ home.htm.
Artificial Intelligence II S. Russell and P. Norvig Artificial Intelligence: A Modern Approach Chapter 11: Planning.
Planning Some material adapted from slides by Tim Finin,Jean-Claude Latombe, Lise Getoor, and Marie desJardins.
CLASSICAL PLANNING AND GRAPHPLAN Classes 17 and 18 All slides created by Dr. Adam P. Anthony.
Planning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 11.
Planning: Part 2 Partial Order Planning COMP151 April 2, 2007.
Artificial Intelligence Chapter 11: Planning
1 Planning Adapted from Tim Finin, UMBC, Marie desJardins Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
1 Lecture 12 example (from slides prepared by Prof. J. Rosenchein)
Planning: Part 1 Representation and State-space Search COMP151 March 30, 2007.
Planning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 11.
Intro to AI Fall 2002 © L. Joskowicz 1 Introduction to Artificial Intelligence LECTURE 12: Planning Motivation Search, theorem proving, and planning Situation.
1 Pertemuan 17 Planning Matakuliah: T0264/Intelijensia Semu Tahun: Juli 2006 Versi: 2/1.
1 Action Planning (Where logic-based representation of knowledge makes search problems more interesting) R&N: Chap. 11, Sect. 11.1–4.
1 Planning Chapters 11 and 12 Thanks: Professor Dan Weld, University of Washington.
AI Principles, Lecture on Planning Planning Jeremy Wyatt.
Planning Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.
PLANNING Partial order regression planning Temporal representation 1 Deductive planning in Logic Temporal representation 2.
An Introduction to Artificial Intelligence CE Chapter 11 – Planning Ramin Halavati In which we see how an agent can take.
Classical Planning Chapter 10.
Feng Zhiyong Tianjin University Fall planning.
1 07. The planning problem 2  Inputs: 1. A description of the world state 2. The goal state description 3. A set of actions  Output: A sequence of actions.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Artificial Intelligence 1: Planning
April 3, 2006AI: Chapter 11: Planning1 Artificial Intelligence Chapter 11: Planning Michael Scherger Department of Computer Science Kent State University.
1 CSC 8520 Spring Paula MatuszekBased on
Planning Chapter Planning The Planning problem Planning with State-space search Partial-order planning Planning graphs Planning with propositional.
Action Planning (Where logic-based representation of knowledge makes search problems more interesting) R&N: Chap. 10.3, Chap. 11, Sect. 11.1–4 (2 nd edition.
CPS 270: Artificial Intelligence Planning Instructor: Vincent Conitzer.
CPS 570: Artificial Intelligence Planning Instructor: Vincent Conitzer.
1 Chapter 16 Planning Methods. 2 Chapter 16 Contents (1) l STRIPS l STRIPS Implementation l Partial Order Planning l The Principle of Least Commitment.
AI Lecture 17 Planning Noémie Elhadad (substituting for Prof. McKeown)
1/16 Planning Chapter 11- Part1 Author: Vali Derhami.
Intelligent Systems Planning.
Planning I: Total Order Planners Sections
April 3, 2006AI: Chapter 11: Planning1 Artificial Intelligence Chapter 11: Planning.
Planning in FOL Systems sequences of actions to achieve goals.
Introduction to Artificial Intelligence April 11, 2012 Planning 1 Introduction to Artificial Intelligence CSCE476/876, Spring 2012:
CS621: Artificial Intelligence Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay Lecture 19: Hidden Markov Models.
Consider the task get milk, bananas, and a cordless drill.
ADVANCED PLANNING TECHNIQUES Dr. Adam Anthony Lecture 22.
Ch. 11 – Planning Supplemental slides for CSE 327 Prof. Jeff Heflin.
1 Planning Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
Introduction to Artificial Intelligence – Unit 6 Planning Course The Hebrew University of Jerusalem School of Engineering and Computer Science Academic.
CLASSICAL PLANNING. Outline  The challenges in planning with standard search algorithm  Representing Plans – the PDDL language  Planning as state -
An Introduction to Artificial Intelligence CE 40417
Planning Chapter 16.
Introduction Contents Sungwook Yoon, Postdoctoral Research Associate
Planning (AIMA Ch. 10) Planning problem defined Simple planning agent
EA C461 – Artificial Intelligence Planning
Consider the task get milk, bananas, and a cordless drill
CS344 : Introduction to Artificial Intelligence
AI Planning.
Logic for Artificial Intelligence
Artificial Intelligence: Planning
Planning Chapter
Planning Problems On(C, A)‏ On(A, Table)‏ On(B, Table)‏ Clear(C)‏
AI Planning Introduction to Artificial Intelligence
CS344 : Introduction to Artificial Intelligence
Russell and Norvig: Chapter 11 CS121 – Winter 2003
CMSC 471 Fall 2002 Class #20/21/23 – Wednesday, November 6 / Monday, November 11 / Monday, November 18.
Prof. Pushpak Bhattacharyya, IIT Bombay
Presentation transcript:

Planning Russell and Norvig: Chapter 11

Planning Agent environment agent ? sensors actuators A1A2A3

Contents Representations in planning Representation of action: preconditions + effects Forward planning Backward chaining

Choose actions to achieve a certain goal But isn’t it exactly the same goal as for problem solving? Some difficulties with problem solving: The successor function is a black box: it must be “applied” to a state to know which actions are possible in that state and what are the effects of each one Goal of Planning

Representations in Planning Planning opens up the black-boxes by using logic to represent: Actions States Goals One possible language: STRIPS Problem solvingLogic representation Planning

State Representation Conjunction of propositions: BLOCK(A), BLOCK(B), BLOCK(C), ON(A,TABLE), ON(B,TABLE), ON(C,A), CLEAR(B), CLEAR(C), HANDEMPTY AB C TABLE

Goal Representation A B C Conjunction of propositions: ON(A,TABLE), ON(B,A), ON(C,B) The goal G is achieved in a state S if all the propositions in G are also in S

Action Representation Unstack(x,y) P = HANDEMPTY, BLOCK(x), BLOCK(y), CLEAR(x), ON(x,y) E =  HANDEMPTY,  CLEAR(x), HOLDING(x),  ON(x,y), CLEAR(y) Precondition: conjunction of propositions Effect: list of literals Means: Remove HANDEMPTY from state Means: Add HOLDING(x) to state

Example AB C Unstack(C,A) P = HANDEMPTY, BLOCK(C), BLOCK(A), CLEAR(C), ON(C,A) E =  HANDEMPTY,  CLEAR(C), HOLDING(C),  ON(C,A), CLEAR(A) BLOCK(A), BLOCK(B), BLOCK(C), ON(A,TABLE), ON(B,TABLE), ON(C,A), CLEAR(B), CLEAR(C), HANDEMPTY

Example AB C BLOCK(A), BLOCK(B), BLOCK(C), ON(A,TABLE), ON(B,TABLE), ON(C,A), CLEAR(B), CLEAR(C), HANDEMPTY, HOLDING(C), CLEAR(A) Unstack(C,A) P = HANDEMPTY, BLOCK(C), BLOCK(A), CLEAR(C), ON(C,A) E =  HANDEMPTY,  CLEAR(C), HOLDING(C),  ON(C,A), CLEAR(A)

Action Representation Unstack(x,y) P = HANDEMPTY, BLOCK(x), BLOCK(y), CLEAR(x), ON(x,y) E =  HANDEMPTY,  CLEAR(x), HOLDING(x),  ON(x,y), CLEAR(y) Stack(x,y) P = HOLDING(x), BLOCK(x), BLOCK(y), CLEAR(y) E = ON(x,y),  CLEAR(y),  HOLDING(x), CLEAR(x), HANDEMPTY PutDown(x) P = HOLDING(x) E = ON(x,TABLE),  HOLDING(x), CLEAR(x), HANDEMPTY Pickup(x) P = HANDEMPTY, BLOCK(x), CLEAR(x), ON(x,TABLE) E =  HANDEMPTY,  CLEAR(x), HOLDING(x),  ON(x,TABLE)

Summary of STRIPS language features Representation of states Decompose the world in logical conditions and represent a state as a conjunction of positive literals. Closed world assumption: conditions not mentioned in state assumed to be false. Representation of goals Partially specified state and represented as a conjunction of positive ground literals A goal is satisfied if the state contains all literals in goal.

Summary of STRIPS language features Representations of actions Action = PRECOND + EFFECT  Action name and parameter list  Precondition (conj. of function-free literals)  Effect (conj of function-free literals) Add-list vs. delete-list in Effect

Semantics The result of executing action a in state s is the state s’ s’ is same as s except  Any positive literal P in the effect of a is added to s’  Any negative literal ¬P is removed from s’ STRIPS assumption: (avoids representational frame problem) every literal NOT in the effect remains unchanged

Expressiveness and extensions STRIPS is simplified Important limit: function-free literals Allows for propositional representation Function symbols lead to infinitely many states and actions Recent extension:Action Description language (ADL)

Example: air cargo transport Init(At(C1, SFO)  At(C2,JFK)  At(P1,SFO)  At(P2,JFK)  Cargo(C1)  Cargo(C2)  Plane(P1)  Plane(P2)  Airport(JFK)  Airport(SFO)) Goal(At(C1,JFK)  At(C2,SFO)) Action(Load(c,p,a) PRECOND: At(c,a)  At(p,a)  Cargo(c)  Plane(p)  Airport(a) EFFECT: ¬At(c,a)  In(c,p)) Action(Unload(c,p,a) PRECOND: In(c,p)  At(p,a)  Cargo(c)  Plane(p)  Airport(a) EFFECT: At(c,a)  ¬In(c,p)) Action(Fly(p,from,to) PRECOND: At(p,from)  Plane(p)  Airport(from)  Airport(to) EFFECT: ¬ At(p,from)  At(p,to)) [Load(C1,P1,SFO), Fly(P1,SFO,JFK), Load(C2,P2,JFK), Fly(P2,JFK,SFO)]

Planning with state-space search Both forward and backward search possible Progression planners forward state-space search Consider the effect of all possible actions in a given state Regression planners backward state-space search To achieve a goal, what must have been true in the previous state.

Progression and regression

Progression algorithm Formulation as state-space search problem: Initial state = initial state of the planning problem  Literals not appearing are false Actions = those whose preconditions are satisfied  Add positive effects, delete negative Goal test = does the state satisfy the goal Step cost = each action costs 1 No functions … any graph search that is complete is a complete planning algorithm. Inefficient: (1) irrelevant action problem (2) good heuristic required for efficient search

Progression (Forward) Planning AB C AB C AB C A C B AC B A C B A C B A B C AB C Unstack(C,A)) Pickup(B) Forward planning searches a space of world states In general, many actions are applicable to a state  huge branching factor

Regression (Backward Chaining) ON(B,A), ON(C,B) Stack(C,B) ON(B,A), HOLDING(C), CLEAR(B) AB C In general, there are much fewer actions relevant to a goal than there are actions applicable to a state  smaller branching factor than with forward planning

Backward Chaining ON(B,A), ON(C,B) Stack(C,B) ON(B,A), HOLDING(C), CLEAR(B) AB C CLEAR(C), ON(C,TABLE), CLEAR(A), HANDEMPTY, CLEAR(B), ON(B,TABLE) CLEAR(C), ON(C,TABLE), HOLDING(B), CLEAR(A) Stack(B,A) Pickup(B) Putdown(C) CLEAR(A), CLEAR(B), ON(B,TABLE), HOLDING(C) Unstack(C,A) CLEAR(B), ON(B,TABLE), CLEAR(C), HANDEMPTY, ON(C,A) Pickup(C) ON(B,A), CLEAR(B), HANDEMPTY, CLEAR(C), ON(C,TABLE) Backward planning searches a space of goals

Regression algorithm How to determine predecessors? What are the states from which applying a given action leads to the goal? Goal state = At(C1, B)  At(C2, B)  …  At(C20, B) Relevant action for first conjunct: Unload(C1,p,B) Works only if pre-conditions are satisfied. Previous state= In(C1, p)  At(p, B)  At(C2, B)  …  At(C20, B) Subgoal At(C1,B) should not be present in this state. Actions must not undo desired literals (consistent) Main advantage: only relevant actions are considered. Often much lower branching factor than forward search.

Heuristics for state-space search Neither progression or regression are very efficient without a good heuristic. How many actions are needed to achieve the goal? Exact solution is NP hard, find a good estimate Two approaches to find admissible heuristic: The optimal solution to the relaxed problem.  Remove all preconditions from actions The subgoal independence assumption: The cost of solving a conjunction of subgoals is approximated by the sum of the costs of solving the subproblems independently.

Partial-order planning Progression and regression planning are totally ordered plan search forms. They cannot take advantage of problem decomposition.  Decisions must be made on how to sequence actions on all the subproblems.

Shoe example Goal(RightShoeOn  LeftShoeOn) Init() Action(RightShoe,PRECOND: RightSockOn EFFECT: RightShoeOn) Action(RightSock,PRECOND: EFFECT: RightSockOn) Action(LeftShoe,PRECOND: LeftSockOn EFFECT: LeftShoeOn) Action(LeftSock,PRECOND: EFFECT: LeftSockOn) Planner: combine two action sequences (1)leftsock, leftshoe (2)rightsock, rightshoe

Partial-order planning Any planning algorithm that can place two actions into a plan without which comes first is a POP.

Summary Representations in planning Representation of action: preconditions + effects Forward planning Backward chaining Partial-order planning