Download presentation
Presentation is loading. Please wait.
Published byDarleen Weaver Modified over 9 years ago
1
Tractable Planning for Real-World Robotics: The promises and challenges of dealing with uncertainty Joelle Pineau Robotics Institute Carnegie Mellon University Stanford University May 3, 2004
2
Joelle PineauTractable Planning for Real-World Robotics2 Robots in unstructured environments
3
Joelle PineauTractable Planning for Real-World Robotics3 A vision for robotic-assisted health-care Moving things around Moving things around Supporting inter-personal communication Supporting inter-personal communication Calling for help in emergencies Calling for help in emergencies Monitoring Rx adherence & safety Monitoring Rx adherence & safety Providing information Providing information Reminding to eat, drink, & take meds Reminding to eat, drink, & take meds Providing physical assistance Providing physical assistance
4
Joelle PineauTractable Planning for Real-World Robotics4 What is hard about planning for real-world robotics? Generating a plan that has high expected utility. Switching between different representations of the world. Resolving conflicts between interfering jobs. Accomplishing jobs in changing, partly unknown environments. Handling percepts which are incomplete, ambiguous, outdated, incorrect. * Highlights from a proposed robot grand challenge
5
Joelle PineauTractable Planning for Real-World Robotics5 Talk outline Uncertainty in plan-based robotics Partially Observable Markov Decision Processes (POMDPs) POMDP solver #1: Point-based value iteration (PBVI) POMDP solver #2: Policy-contingent abstraction (PolCA+)
6
Joelle PineauTractable Planning for Real-World Robotics6 Why use a POMDP? POMDPs provide a rich framework for sequential decision-making, which can model: –Effect uncertainty –State uncertainty –Varying rewards across actions and goals
7
Joelle PineauTractable Planning for Real-World Robotics7 Robotics: Robust mobile robot navigation [Simmons & Koenig, 1995; + many more] Autonomous helicopter control [Bagnell & Schneider, 2001; Ng et al., 2003] Machine vision [Bandera et al., 1996; Darrell & Pentland, 1996] High-level robot control [Pineau et al., 2003] Robust dialogue management [Roy, Pineau & Thrun, 2000; Peak & Horvitz, 2000] POMDP applications in last decade Others: Machine maintenance [Puterman., 1994] Network troubleshooting [Thiebeaux et al., 1996] Circuits testing [correspondence, 2004] Preference elicitation [Boutilier, 2002] Medical diagnosis [Hauskrecht, 1997]
8
Joelle PineauTractable Planning for Real-World Robotics8 POMDP model POMDP is n-tuple { S, A, Z, T, O, R }: What goes on: s t-1 stst a t-1 atat S = state set A = action set Z = observation set What we see: z t-1 ztzt
9
Joelle PineauTractable Planning for Real-World Robotics9 POMDP model POMDP is n-tuple { S, A, Z, T, O, R }: What goes on: s t-1 stst a t-1 atat T: Pr(s’|s,a) = state-to-state transition probabilitiesS = state set A = action set Z = observation set What we see: z t-1 ztzt
10
Joelle PineauTractable Planning for Real-World Robotics10 POMDP model POMDP is n-tuple { S, A, Z, T, O, R }: What goes on: s t-1 stst a t-1 atat T: Pr(s’|s,a) = state-to-state transition probabilities O: Pr(z|s,a) = observation generation probabilities S = state set A = action set Z = observation set What we see: z t-1 ztzt
11
Joelle PineauTractable Planning for Real-World Robotics11 POMDP model POMDP is n-tuple { S, A, Z, T, O, R }: What goes on: s t-1 stst a t-1 atat T: Pr(s’|s,a) = state-to-state transition probabilities O: Pr(z|s,a) = observation generation probabilities R(s,a) = reward function S = state set A = action set Z = observation set What we see: z t-1 ztzt r t-1 rtrt
12
Joelle PineauTractable Planning for Real-World Robotics12 POMDP model POMDP is n-tuple { S, A, Z, T, O, R }: What goes on: s t-1 stst a t-1 atat What we see: z t-1 ztzt What we infer: b t-1 btbt r t-1 rtrt T: Pr(s’|s,a) = state-to-state transition probabilities O: Pr(z|s,a) = observation generation probabilities R(s,a) = reward function S = state set A = action set Z = observation set
13
Joelle PineauTractable Planning for Real-World Robotics13 Examples of robot beliefs robot particles Uniform belief Bi-modal belief
14
Joelle PineauTractable Planning for Real-World Robotics14 Understanding the belief state A belief is a probability distribution over states Where Dim(B) = |S|-1 –E.g. Let S={s 1, s 2 } P(s 1 ) 0 1
15
Joelle PineauTractable Planning for Real-World Robotics15 Understanding the belief state A belief is a probability distribution over states Where Dim(B) = |S|-1 –E.g. Let S={s 1, s 2, s 3 } P(s 1 ) P(s 2 ) 0 1 1
16
Joelle PineauTractable Planning for Real-World Robotics16 Understanding the belief state A belief is a probability distribution over states Where Dim(B) = |S|-1 –E.g. Let S={s 1, s 2, s 3, s 4 } P(s 1 ) P(s 2 ) 0 1 1 P(s 3 )
17
Joelle PineauTractable Planning for Real-World Robotics17 POMDP solving Objective: Find the sequence of actions that maximizes the expected sum of rewards. Value function Immediate reward Future reward
18
Joelle PineauTractable Planning for Real-World Robotics18 Represent V(b) as the upper surface of a set of vectors. –Each vector is a piece of the control policy (= action sequences). –Dim(vector) = number of states. Modify / add vectors to update value fn (i.e. refine policy). POMDP value function P(s 1 ) V(b) b 2 states
19
Joelle PineauTractable Planning for Real-World Robotics19 Optimal POMDP solving Simple problem: 2 states, 3 actions, 3 observations V 0 (b) b Plan length #vectors 0 1 P(break-in)
20
Joelle PineauTractable Planning for Real-World Robotics20 Optimal POMDP solving Simple problem: 2 states, 3 actions, 3 observations P(break-in) V 1 (b) b Plan length # vectors 0 1 1 3 Call-911 Investigate Go-to-bed
21
Joelle PineauTractable Planning for Real-World Robotics21 Optimal POMDP solving Simple problem: 2 states, 3 actions, 3 observations V 2 (b) b Plan length # vectors 0 1 1 3 2 27 P(break-in) Call-911 Investigate Go-to-bed
22
Joelle PineauTractable Planning for Real-World Robotics22 Optimal POMDP solving Simple problem: 2 states, 3 actions, 3 observations V 3 (b) b Plan length # vectors 0 1 1 3 2 27 3 2187 P(break-in) Call-911 Investigate Go-to-bed
23
Joelle PineauTractable Planning for Real-World Robotics23 Optimal POMDP solving Simple problem: 2 states, 3 actions, 3 observations Plan length # vectors 0 1 1 3 2 27 3 2187 4 14,348,907 V 4 (b) b P(break-in) Call-911 Investigate Go-to-bed
24
Joelle PineauTractable Planning for Real-World Robotics24 The curse of history Policy size grows exponentially with the planning horizon: WhereΓ = policy size n = planning horizon A = # actions Z = # observations
25
Joelle PineauTractable Planning for Real-World Robotics25 How many vectors for this problem? 10 4 (navigation) x 10 3 (dialogue) states 1000+ observations 100+ actions
26
Joelle PineauTractable Planning for Real-World Robotics26 Talk outline Uncertainty in plan-based robotics Partially Observable Markov Decision Processes (POMDPs) POMDP solver #1: Point-based value iteration (PBVI) POMDP solver #2: Policy-contingent abstraction (PolCA+)
27
Joelle PineauTractable Planning for Real-World Robotics27 Exact solving assumes all beliefs are equally likely robot particles Uniform beliefBi-modal beliefN-modal belief INSIGHT: No sequence of actions and observations can produce this N-modal belief.
28
Joelle PineauTractable Planning for Real-World Robotics28 A new algorithm: Point-based value iteration P(s 1 ) V(b) b1b1 b0b0 b2b2 Approach: Select a small set of belief points Plan for those belief points only
29
Joelle PineauTractable Planning for Real-World Robotics29 A new algorithm: Point-based value iteration P(s 1 ) V(b) b1b1 b0b0 b2b2 a,z Approach: Select a small set of belief points Use well-separated, reachable beliefs Plan for those belief points only
30
Joelle PineauTractable Planning for Real-World Robotics30 A new algorithm: Point-based value iteration P(s 1 ) V(b) b1b1 b0b0 b2b2 a,z Approach: Select a small set of belief points Use well-separated, reachable beliefs Plan for those belief points only Learn value and its gradient
31
Joelle PineauTractable Planning for Real-World Robotics31 A new algorithm: Point-based value iteration Approach: Select a small set of belief points Use well-separated, reachable beliefs Plan for those belief points only Learn value and its gradient Pick action that maximizes value: P(s 1 ) V(b) b b1b1 b0b0 b2b2
32
Joelle PineauTractable Planning for Real-World Robotics32 The anytime PBVI algorithm Alternate between: 1. Growing the set of belief point 2. Planning for those belief points Terminate when you run out of time or have a good policy.
33
Joelle PineauTractable Planning for Real-World Robotics33 Complexity of value update Exact UpdatePBVI Time: ProjectionS 2 A Z n S 2 A Z n SumS A n Z S A Z n B Size: (# vectors) A n Z B where:S = # states n = # vectors at iteration n A = # actionsB = # belief points Z = # observations
34
Joelle PineauTractable Planning for Real-World Robotics34 Theoretical properties of PBVI Theorem: For any set of belief points B and planning horizon n, the error of the PBVI algorithm is bounded by: P(s 1 ) V(b) b1b1 b0b0 b2b2 Where is the set of reachable beliefs B is the set of all beliefs Err
35
Joelle PineauTractable Planning for Real-World Robotics35 The anytime PBVI algorithm Alternate between: 1. Growing the set of belief point 2. Planning for those belief points Terminate when you run out of time or have a good policy.
36
Joelle PineauTractable Planning for Real-World Robotics36 PBVI’s belief selection heuristic 1. Leverage insight from policy search methods: –Focus on reachable beliefs. P(s 1 ) bb a1,z2 b a2,z2 b a2,z1 b a1,z1 a2,z2a1,z2 a2,z1 a1,z1
37
Joelle PineauTractable Planning for Real-World Robotics37 PBVI’s belief selection heuristic 1. Leverage insight from policy search methods: –Focus on reachable beliefs. 2. Focus on high probability beliefs: –Consider all actions, but stochastic observation choice. P(s 1 ) bb a1,z2 b a2,z1 a1,z2 a2,z1 b a2,z2 b a1,z1
38
Joelle PineauTractable Planning for Real-World Robotics38 PBVI’s belief selection heuristic 1. Leverage insight from policy search methods: –Focus on reachable beliefs. 2. Focus on high probability beliefs: –Consider all actions, but stochastic observation choice. 3. Use the error bound on point-based value updates: –Select well-separated beliefs, rather than near-by beliefs. P(s 1 ) bb a1,z2 b a2,z1 a1,z2 a2,z1
39
Joelle PineauTractable Planning for Real-World Robotics39 Classes of value function approximations 1. No belief [Littman et al., 1995] 3. Compressed belief [Poupart&Boutilier, 2002; Roy&Gordon, 2002] x1x1 x0x0 x2x2 2. Grid over belief [Lovejoy, 1991; Brafman 1997; Hauskrecht, 2000; Zhou&Hansen, 2001] 4. Sample belief points [Poon, 2001; Pineau et al, 2003]
40
Joelle PineauTractable Planning for Real-World Robotics40 Performance on well-known POMDPs Maze1 0.20 0.94 0.00 2.30 2.25 REWARD Maze2 0.11 - 0.07 0.35 0.34 Maze3 0.26 - 0.11 0.53 Maze1 0.19 - 24hrs 12166 3448 TIME Maze2 1.44 - 24hrs 27898 360 Maze3 0.51 - 24hrs 450 288 Maze1 - 174 - 660 470 # Belief points Maze2 - 337 - 1840 95 Maze3 - 300 86 Method No belief [Littman&al., 1995] Grid [Brafman., 1997] Compressed [Poupart&al., 2003] Sample [Poon, 2001] PBVI [Pineau&al., 2003] Maze1: 36 states Maze2: 92 states Maze3: 60 states
41
Joelle PineauTractable Planning for Real-World Robotics41 PBVI in the Nursebot domain Objective: Find the patient! State space = RobotPosition PatientPosition Observation space = RobotPosition + PatientFound Action space = {North, South, East, West, Declare}
42
Joelle PineauTractable Planning for Real-World Robotics42 PBVI performance on find-the-patient domain Patient found 17% of trials Patient found 90% of trials No Belief PBVI No Belief PBVI
43
Joelle PineauTractable Planning for Real-World Robotics43 Validation of PBVI’s belief expansion heuristic Greedy PBVI No Belief Random Find-the-patient domain 870 states, 5 actions, 30 observations
44
Joelle PineauTractable Planning for Real-World Robotics44 Policy assuming full observability You find some: (25%) You loose some: (75%)
45
Joelle PineauTractable Planning for Real-World Robotics45 PBVI Policy with 3141 belief points You find some: (81%) You loose some: (19%)
46
Joelle PineauTractable Planning for Real-World Robotics46 PBVI Policy with 643 belief points You find some: (22%) You loose some: (78%)
47
Joelle PineauTractable Planning for Real-World Robotics47 Highlights of the PBVI algorithm Algorithmic : –New belief sampling algorithm. –Efficient heuristic for belief point selection. –Anytime performance. Experimental : –Outperforms previous value approximation algorithms on known problems. –Solves new larger problem (1 order of magnitude increase in problem size). Theoretical : –Bounded approximation error. [ Pineau, Gordon & Thrun, IJCAI 2003. Pineau, Gordon & Thrun, NIPS 2003. ]
48
Joelle PineauTractable Planning for Real-World Robotics48 Back to the grand challenge How can we go from 10 3 states to real-world problems?
49
Joelle PineauTractable Planning for Real-World Robotics49 Talk outline Uncertainty in plan-based robotics Partially Observable Markov Decision Processes (POMDPs) POMDP solver #1: Point-based value iteration (PBVI) POMDP solver #2: Policy-contingent abstraction (PolCA+)
50
Joelle PineauTractable Planning for Real-World Robotics50 Navigation Structured POMDPs Many real-world decision-making problems exhibit structure inherent to the problem domain. Cognitive supportSocial interaction High-level controller Move AskWhere LeftRightForwardBackward
51
Joelle PineauTractable Planning for Real-World Robotics51 Structured POMDP approaches Factored models [Boutilier & Poole, 1996; Hansen & Feng, 2000; Guestrin et al., 2001] –Idea: Represent state space with multi-valued state features. –Insight: Independencies between state features can be leveraged to overcome the curse of dimensionality. Hierarchical POMDPs [Wiering & Schmidhuber, 1997; Theocharous et al., 2000; Hernandez-Gardiol & Mahadevan, 2000; Pineau & Thrun, 2000] –Idea: Exploit domain knowledge to divide one POMDP into many smaller ones. –Insight: Smaller action sets further help overcome the curse of history.
52
Joelle PineauTractable Planning for Real-World Robotics52 A hierarchy of POMDPs Act ExamineHealth Navigate Move VerifyFluids ClarifyGoal NorthSouthEastWest VerifyMeds subtask abstract action primitive action
53
Joelle PineauTractable Planning for Real-World Robotics53 PolCA+: Planning with a hierarchy of POMDPs Navigate Move ClarifyGoal SouthEast West North A Move = {N,S,E,W} ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds Step 1: Select the action set
54
Joelle PineauTractable Planning for Real-World Robotics54 PolCA+: Planning with a hierarchy of POMDPs Navigate Move ClarifyGoal SouthEast West North A Move = {N,S,E,W} S Move = {s 1,s 2 } STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds Step 1: Select the action set Step 2: Minimize the state set
55
Joelle PineauTractable Planning for Real-World Robotics55 PolCA+: Planning with a hierarchy of POMDPs Navigate Move ClarifyGoal SouthEast West North A Move = {N,S,E,W} S Move = {s 1,s 2 } STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds PARAMETERS {b h,T h,O h,R h } PARAMETERS {b h,T h,O h,R h } Step 1: Select the action set Step 2: Minimize the state set Step 3: Choose parameters
56
Joelle PineauTractable Planning for Real-World Robotics56 PolCA+: Planning with a hierarchy of POMDPs Navigate Move ClarifyGoal SouthEast West North STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus STATE FEATURES X-position Y-position X-goal Y-goal HealthStatus ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds ACTIONS North South East West ClarifyGoal VerifyFluids VerifyMeds PLAN h PLAN h PARAMETERS {b h,T h,O h,R h } PARAMETERS {b h,T h,O h,R h } Step 1: Select the action set Step 2: Minimize the state set Step 3: Choose parameters Step 4: Plan task h A Move = {N,S,E,W} S Move = {s 1,s 2 }
57
Joelle PineauTractable Planning for Real-World Robotics57 PolCA+ in the Nursebot domain Goal: A robot is deployed in a nursing home, where it provides reminders to elderly users and accompanies them to appointments.
58
Joelle PineauTractable Planning for Real-World Robotics58 Performance measure Hierarchy + Belief Execution Steps Hierarchy + Belief PolCA+
59
Joelle PineauTractable Planning for Real-World Robotics59 Comparing user performance 0.1 0.18 POMDP Policy No Belief Policy
60
Joelle PineauTractable Planning for Real-World Robotics60 Visit to the nursing home
61
Joelle PineauTractable Planning for Real-World Robotics61 Highlights of the PolCA+ algorithm Algorithmic: –New hierarchical approach for POMDP framework. –POMDP-specific state and observation abstraction methods. Experimental: –First instance of POMDP-based high-level robot controller. –Novel application of POMDPs to robust dialogue management. Theoretical: –For special case (fully observable), guarantees recursive optimality. [ Pineau, Gordon & Thrun, UAI 2003. Pineau et al., RAS 2003. Roy, Pineau & Thrun, ACL 2001]
62
Joelle PineauTractable Planning for Real-World Robotics62 Future work PBVI: How can we handle domains with multi-valued state features? Can we leverage dimensionality reduction? Can we find better ways to pick belief points? PolCA+: Can we automatically learn hierarchies? How can we learn (or do without) pseudo-reward functions?
63
Joelle PineauTractable Planning for Real-World Robotics63 Questions? Project information: www.cs.cmu.edu/~nursebot Navigation software: www.cs.cmu.edu/~carmen Papers and more: www.cs.cmu.edu/~jpineau Collaborators: Geoffrey Gordon, Judith Matthews, Michael Montemerlo, Martha Pollack, Nicholas Roy, Sebastian Thrunh
64
Joelle PineauTractable Planning for Real-World Robotics64 Two types of uncertainty Effect→ Stochastic action effects State→ Partial and noisy sensor information
65
Joelle PineauTractable Planning for Real-World Robotics65 Example: Effect uncertainty Start position Distribution over possible next-step positions Start position Distribution over possible next-step positions Motion action
66
Joelle PineauTractable Planning for Real-World Robotics66 Two types of uncertainty Effect→ Stochastic action effects State→ Partial and noisy sensor information
67
Joelle PineauTractable Planning for Real-World Robotics67 Example: State uncertainty Effect → Stochastic action effects State → Partial and noisy sensor information Model → Inaccurate parameterization of the environment Agent → Unknown behaviour of other agents Start position Distribution over possible next-step positions
68
Joelle PineauTractable Planning for Real-World Robotics68 Validation of PBVI’s belief expansion heuristic No Belief Hallway domain 60 states, 5 actions, 20 observations
69
Joelle PineauTractable Planning for Real-World Robotics69 State space reduction No hierarchy PolCA+
70
Joelle PineauTractable Planning for Real-World Robotics70 Future directions Improving POMDP planning –sparser belief space sampling, ordered value updating, dimensionality reduction, continuous / hybrid domains Addressing two more types of uncertainty: 1. Effect 2. State 3. Model 4. Agent Exploring new applications
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.