Extending PDDL to Model Stochastic Decision Processes Håkan L. S. Younes Carnegie Mellon University.

Slides:



Advertisements
Similar presentations
Computer Science CPSC 322 Lecture 25 Top Down Proof Procedure (Ch 5.2.2)
Advertisements

Causal-link Planning II José Luis Ambite. 2 CS 541 Causal Link Planning II Planning as Search State SpacePlan Space AlgorithmProgression, Regression POP.
Distributed Markov Chains P S Thiagarajan School of Computing, National University of Singapore Joint work with Madhavan Mukund, Sumit K Jha and Ratul.
Modified MDPs for Concurrent Execution AnYuan Guo Victor Lesser University of Massachusetts.
Situation Calculus for Action Descriptions We talked about STRIPS representations for actions. Another common representation is called the Situation Calculus.
Time Constraints in Planning Sudhan Kanitkar
Plan Generation & Causal-Link Planning 1 José Luis Ambite.
Hybrid Systems Presented by: Arnab De Anand S. An Intuitive Introduction to Hybrid Systems Discrete program with an analog environment. What does it mean?
Probabilistic Verification of Discrete Event Systems using Acceptance Sampling Håkan L. S. YounesReid G. Simmons Carnegie Mellon University.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Department of Computer Science Undergraduate Events More
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
CSE-573 Artificial Intelligence Partially-Observable MDPS (POMDPs)
Modeling & Simulation. System Models and Simulation Framework for Modeling and Simulation The framework defines the entities and their Relationships that.
A Brief Introduction. Acknowledgements  The material in this tutorial is based in part on: Concurrency: State Models & Java Programming, by Jeff Magee.
IPC 2004 Probabilistic Planning Track Michael L. LittmanHåkan L. S. Younes Rutgers UniversityCarnegie Mellon University.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Statistical Probabilistic Model Checking Håkan L. S. Younes Carnegie Mellon University.
Ymer: A Statistical Model Checker Håkan L. S. Younes Carnegie Mellon University.
Probabilistic Verification of Discrete Event Systems Håkan L. S. Younes Reid G. Simmons (initial work performed at HTC, Summer 2001)
1 Chapter 16 Planning Methods. 2 Chapter 16 Contents (1) l STRIPS l STRIPS Implementation l Partial Order Planning l The Principle of Least Commitment.
Decision Making: An Introduction 1. 2 Decision Making Decision Making is a process of choosing among two or more alternative courses of action for the.
TCOM 501: Networking Theory & Fundamentals
Planning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 11.
Temporal Specification Chris Patel Vinay Viswanathan.
Markov Decision Processes CSE 473 May 28, 2004 AI textbook : Sections Russel and Norvig Decision-Theoretic Planning: Structural Assumptions.
Planning CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanning Supervised.
Performance Analysis of Chi Models using Discrete-Time Probabilistic Reward Graphs N. Trčka, S. Georgievska, J. Markovski, S. Andova, and E.P. de Vink.
Probabilistic Verification of Discrete Event Systems Håkan L. S. Younes.
Probabilistic Temporal Planning with Uncertain Durations Mausam Joint work with Daniel S. Weld University of Washington Seattle.
1 Formal Engineering of Reliable Software LASER 2004 school Tutorial, Lecture1 Natasha Sharygina Carnegie Mellon University.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
A Framework for Planning in Continuous-time Stochastic Domains Håkan L. S. Younes Carnegie Mellon University David J. MuslinerReid G. Simmons Honeywell.
Policy Generation for Continuous-time Stochastic Domains with Concurrency Håkan L. S. YounesReid G. Simmons Carnegie Mellon University.
Planning with Concurrency in Continuous-time Stochastic Domains (thesis proposal) Håkan L. S. Younes Thesis Committee: Reid G. Simmons (chair) Geoffrey.
Stochastic Dynamic Programming with Factored Representations Presentation by Dafna Shahaf (Boutilier, Dearden, Goldszmidt 2000)
(C) 2009 J. M. Garrido1 Object Oriented Simulation with Java.
1 Endgame Logistics  Final Project Presentations  Tuesday, March 19, 3-5, KEC2057  Powerpoint suggested ( to me before class)  Can use your own.
1 Performance Evaluation of Computer Networks: Part II Objectives r Simulation Modeling r Classification of Simulation Modeling r Discrete-Event Simulation.
Planning and Verification for Stochastic Processes with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
CSE-473 Artificial Intelligence Partially-Observable MDPS (POMDPs)
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Planning and Execution with Phase Transitions Håkan L. S. Younes Carnegie Mellon University Follow-up paper to Younes & Simmons’ “Solving Generalized Semi-Markov.
Homework 1 ( Written Portion )  Max : 75  Min : 38  Avg : 57.6  Median : 58 (77%)
CIS 540 Principles of Embedded Computation Spring Instructor: Rajeev Alur
CSE-573 Reinforcement Learning POMDPs. Planning What action next? PerceptsActions Environment Static vs. Dynamic Fully vs. Partially Observable Perfect.
1 Elements of Queuing Theory The queuing model –Core components; –Notation; –Parameters and performance measures –Characteristics; Markov Process –Discrete-time.
Pattern-directed inference systems
Haley: A Hierarchical Framework for Logical Composition of Web Services Haibo Zhao, Prashant Doshi LSDIS Lab, Dept. of Computer Science, University of.
By: Messias, Spaan, Lima Presented by: Mike Plasker DMES – Ocean Engineering.
Carnegie Mellon University
Decision-Theoretic Planning with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Probabilistic Verification of Discrete Event Systems using Acceptance Sampling Håkan L. S. Younes Carnegie Mellon University.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Probabilistic Verification of Discrete Event Systems Håkan L. S. Younes.
A Formalism for Stochastic Decision Processes with Asynchronous Events Håkan L. S. YounesReid G. Simmons Carnegie Mellon University.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Håkan L. S. YounesDavid J. Musliner Carnegie Mellon UniversityHoneywell Laboratories Probabilistic Plan Verification through Acceptance Sampling.
State Machine Model.
Modeling and Simulation (An Introduction)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Review for the Midterm Exam
Markov Decision Processes
Markov Decision Processes
Probability and Time: Markov Models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Probability and Time: Markov Models
Reinforcement Learning Dealing with Partial Observability
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Presentation transcript:

Extending PDDL to Model Stochastic Decision Processes Håkan L. S. Younes Carnegie Mellon University

Introduction PDDL extensions for modeling stochastic decision processes Only full observability is considered Formalism for expressing probabilistic temporally extended goals No commitment on plan representation

Simple Example (:action flip-coin :parameters (?coin) :precondition (holding ?coin) :effect (and(not (holding ?coin)) (probabilistic0.5 (head-up ?coin) 0.5 (tail-up ?coin))))

Stochastic Actions Variation of factored probabilistic STRIPS operators [Dearden & Boutilier 97] An action consists of a precondition  and a consequence set C = {c 1, …, c n } Each c i has a trigger condition  i and an effects list E i =  p 1 i, E 1 i ; …; p k i, E k i   j p j = 1 for each E i

Stochastic Actions: Semantics An action is enabled in a state s if its precondition  holds in s Executing a disabled action is allowed, but does not change the state Different from deterministic PDDL Motivation: partial observability Precondition becomes factored trigger condition

Stochastic Actions: Semantics (cont.) When applying an enabled action to s: Select an effect set for each consequence with enabled trigger condition The combined effects of the selected effect sets are applied atomically to s Unique next state if consequences with mutually consistent trigger conditions have commutative effect sets

Syntax of Probabilistic Effects ::= ::= (and *) ::= (forall ( ) ) ::= (when ) ::= (probabilistic +) ::= ::= (and *) ::= ::= (not ) ::= ::= ( ) ::= Any rational number in the interval [0, 1]

Correspondence to Components of Stochastic Actions Effects list: (probabilistic p 1 i E 1 i … p k i E k i ) Consequence: (when  (probabilistic p 1 i E 1 i … p k i E k i ))

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)))) (when (not (office)) (probabilistic 0.9 (office))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)) 0.1 (and))) (when (not (office)) (probabilistic 0.9 (office) 0.1 (and))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet) 0.1 (and))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)) 0.1 (and))) (when (not (office)) (probabilistic 0.9 (office) 0.1 (and))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet) 0.1 (and))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)) 0.1 (and))) (when (not (office)) (probabilistic 0.9 (office) 0.1 (and))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet) 0.1 (and))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)) 0.1 (and))) (when (not (office)) (probabilistic 0.9 (office) 0.1 (and))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet) 0.1 (and))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Stochastic Actions: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)) 0.1 (and))) (when (not (office)) (probabilistic 0.9 (office) 0.1 (and))) (when (and (rain) (not (umbrella))) (probabilistic 0.9 (wet) 0.1 (and))))) office rain   umbrella  wet  office rain   umbrella  wet office rain   umbrella wet  office rain   umbrella wet office  rain  umbrella  wet  office  rain  umbrella  wet office  rain  umbrella wet  office  rain  umbrella wet

Exogenous Events Like stochastic actions, but beyond the control of the decision maker Defined using :event keyword instead of :action keyword Common in control theory to say that everything is an event, and that some are controllable (what we call actions)

Exogenous Events: Example (:action move :parameters () :effect (and (when (office) (probabilistic 0.9 (not (office)))) (when (not (office)) (probabilistic 0.9 (office))))) (:event make-wet :parameters () :precondition (and (rain) (not (umbrella))) :effect (probabilistic 0.9 (wet)))

Expressiveness Discrete-time MDPs Exogenous events are so far only a modeling convenience and do not add to the expressiveness

Adding Time States have stochastic duration Transitions are instantaneous

Actions and Events with Stochastic Delay Associate a delay distribution F(t) with each action a F(t) is the cumulative distribution function for the delay from when a is enabled until it triggers Analogous for exogenous events

Delayed actions and events: Example (:delayed-action move :parameters () :delay (geometric 0.9) :effect (and (when (office) (not (office))) (when (not (office)) (office)))) office  office G(0.9)

Delayed actions and events: Example (:delayed-action move :parameters () :delay (geometric 0.9) :effect (and (when (office) (not (office))) (when (not (office)) (office)))) office  office

Expressiveness Geometric delay distributions Discrete-time MDP Exponential delay distributions Continuous-time MDP General delay distributions At least semi-Markov decision process

General Delay Distribution: Example (:delayed-action move :parameters () :delay (uniform 0 6) :effect (and (when (office) (not (office))) (when (not (office)) (office)))) (:delayed-event make-wet :parameters () :delay (weibull 2) :precondition (and (rain) (not (umbrella))) :effect (wet))

Concurrent Semi-Markov Processes Each action and event separately office  office U(0, 6)  wet wet W(2) move make-wet

Generalized Semi-Markov Process Putting it together office  wet  office  wet U(0, 6) office wet  office wet W(2) U(0, 6) W(2)

Generalized Semi-Markov Process (cont.) Why generalized? office  wet  office  wet U(0, 6) office wet  office wet W(2) U(0, 6) W(2)  office  wet t=0

Generalized Semi-Markov Process (cont.) Why generalized? office  wet  office  wet U(0, 6) office wet  office wet W(2) U(0, 6) U(0, 3) W(2)  office  wet t=0  office wet t=3 make-wet

Expressiveness Hierarchy of stochastic decision processes GSMDP SMDP MDP memoryless delay distributions probabilistic effects general delay distributions probabilistic effects concurrency general delay distributions probabilistic effects

Probabilistic Temporally Extended Goals Goal specified as a CSL (PCTL) formula  ::= true | a |    |  | Pr  p (  )  ::=  U ≤t  | ◊ ≤t  | □ ≤t 

Goals: Examples Achievement goal Pr ≥0.9 (◊ office) Achievement goal with deadline Pr ≥0.9 (◊ ≤5 office) Achievement goal with safety constraint Pr ≥0.9 (  wet U office) Maintenance goal Pr ≥0.8 (□  wet)

Summary PDDL Extensions Probabilistic effects Exogenous events Delayed actions and events CSL/PCTL goals Thursday: Planning!

Panel Statement Stochastic decision processes are useful Robotic control, queuing systems, … Baseline PDDL should be designed with stochastic decision processes in mind Formalisms are needed to express complex constrains on valid plans PCTL/CSL goals

Role of PDDL discrete dynamics continuous dynamicsstochastic dynamics