Download presentation
Presentation is loading. Please wait.
Published byMalakai Fenley Modified over 10 years ago
1
Online Sampling for Markov Decision Processes Bob Givan Joint work w/ E. K. P. Chong, H. Chang, G. Wu Electrical and Computer Engineering Purdue University
2
Bob Givan Electrical and Computer Engineering Purdue University 1 November 4-9, 2001 Markov Decision Process (MDP) Ingredients: System state x in state space X Control action a in A(x) Reward R(x,a) State-transition probability P(x,y,a) Find control policy to maximize objective fun
3
Bob Givan Electrical and Computer Engineering Purdue University 2 November 4-9, 2001 Optimal Policies Policy – mapping from state and time to actions Stationary Policy – mapping from state to actions Goal – a policy maximizing the objective function V H *(x 0 ) = max Obj [R(x 0,a 0 ), …, R(x H-1,a H-1 )] where the “max” is over all policies u = u 0,…,u H-1 For large H, a 0 independent of H. (w/ergodicity assum.) Stationary optimal action a 0 for H = via receding horizon control
4
Bob Givan Electrical and Computer Engineering Purdue University 3 November 4-9, 2001 Q Values Fix a large H, focus on finite-horizon reward Define Q(x,a) = R(x,a) + E[V H-1 *(y)] “Utility” of action a at state x. Name: Q-value of action a at state x. Key identities (Bellman’s equations): V H *(x) = max a Q(x,a) 0 * (x) = argmax a Q(x,a)
5
Bob Givan Electrical and Computer Engineering Purdue University 4 November 4-9, 2001 Solution Methods Recall: u 0 * (x) = argmax a Q(x,a) Q(x,a) = R(x,a) + E [V H-1 *(y)] Problem: Q-value depends on optimal policy. State space is extremely large (often continuous) Two-pronged solution approach: Apply a receding-horizon method Estimate Q-values via simulation/sampling
6
Bob Givan Electrical and Computer Engineering Purdue University 5 November 4-9, 2001 Methods for Q-value Estimation Previous work by other authors: Unbiased sampling (exact Q value) [Kearns et al., IJCAI-99] Policy rollout (lower bound) [Bertsekas & Castanon, 1999] Our techniques: Hindsight optimization (upper bound) Parallel rollout (lower bound)
7
Bob Givan Electrical and Computer Engineering Purdue University 6 November 4-9, 2001 Expectimax Tree for V *
8
Bob Givan Electrical and Computer Engineering Purdue University 7 November 4-9, 2001 Unbiased Sampling
9
Bob Givan Electrical and Computer Engineering Purdue University 8 November 4-9, 2001 Unbiased Sampling (Cont’d) For a given desired accuracy, how large should sampling width and depth be? Answered: Kearns, Mansour, and Ng (1999) Requires prohibitive sampling width and depth e.g. C 10 8, H s > 60 to distinguish “best” and “worst” policies in our scheduling domain We evaluate with smaller width and depth
10
Bob Givan Electrical and Computer Engineering Purdue University 9 November 4-9, 2001 How to Look Deeper?
11
Bob Givan Electrical and Computer Engineering Purdue University 10 November 4-9, 2001 Policy Roll-out
12
Bob Givan Electrical and Computer Engineering Purdue University 11 November 4-9, 2001 Policy Rollout in Equations Write V H u (y) for the value of following policy u Recall: Q(x,a)= R(x,a) + E [V H-1 *(y)] = R(x,a) + E [max u V H-1 u (y)] Given a base policy u, use R(x,a) + E [V H-1 u (y)] as an lower bound estimate of Q-value. Resulting policy is PI(u), given infinite sampling
13
Bob Givan Electrical and Computer Engineering Purdue University 12 November 4-9, 2001 Policy Roll-out (cont’d)
14
Bob Givan Electrical and Computer Engineering Purdue University 13 November 4-9, 2001 Parallel Policy Rollout Generalization of policy rollout, due to [Chang, Givan, and Chong, 2000] Given a set U of base policies, use R(x,a) + E [max u ∊ U V H-1 u (y)] as an estimate of Q-value More accurate estimate than policy rollout Still gives a lower bound to true Q-value Still gives a policy no worse than any in U
15
Bob Givan Electrical and Computer Engineering Purdue University 14 November 4-9, 2001 Hindsight Optimization – Tree View
16
Bob Givan Electrical and Computer Engineering Purdue University 15 November 4-9, 2001 Hindsight Optimization – Equations Swap Max and Exp in expectimax tree. Solve each off-line optimization problem O (kC’ f(H)) time where f(H) is the offline problem complexity Jensen’s inequality implies upper bounds
17
Bob Givan Electrical and Computer Engineering Purdue University 16 November 4-9, 2001 Hindsight Optimization (Cont’d)
18
Bob Givan Electrical and Computer Engineering Purdue University 17 November 4-9, 2001 Application to Example Problems Apply unbiased sampling, policy rollout, parallel rollout, and hindsight optimization to: Multi-class deadline scheduling Random early dropping Congestion control
19
Bob Givan Electrical and Computer Engineering Purdue University 18 November 4-9, 2001 Basic Approach Traffic model provides a stochastic description of possible future outcomes Method Formulate network decision problems as POMDPs by incorporating traffic model Solve belief-state MDP online using sampling (choose time-scale to allow for computation time)
20
Bob Givan Electrical and Computer Engineering Purdue University 19 November 4-9, 2001 Domain 1: Deadline Scheduling Objective: Minimize weighted loss
21
Bob Givan Electrical and Computer Engineering Purdue University 20 November 4-9, 2001 Domain 2: Random Early Dropping Objective: Minimize delay without sacrificing throughput
22
Bob Givan Electrical and Computer Engineering Purdue University 21 November 4-9, 2001 Domain 3: Congestion Control
23
Bob Givan Electrical and Computer Engineering Purdue University 22 November 4-9, 2001 Traffic Modeling A Hidden Markov Model (HMM) for each source Note: state is hidden, model is partially observed
24
Bob Givan Electrical and Computer Engineering Purdue University 23 November 4-9, 2001 Deadline Scheduling Results Non-sampling Policies: EDF: earliest deadline first. Deadline sensitive, class insensitive. SP: static priority. Deadline insensitive, class sensitive. CM: current minloss [Givan et al., 2000] Deadline and class sensitive. Minimizes weighted loss for the current packets.
25
Bob Givan Electrical and Computer Engineering Purdue University 24 November 4-9, 2001 Deadline Scheduling Results Objective: minimize weighted loss Comparison: Non-sampling policies Unbiased sampling (Kearns et al.) Hindsight optimization Rollout with CM as base policy Parallel rollout Results due to H. S. Chang
26
Bob Givan Electrical and Computer Engineering Purdue University 25 November 4-9, 2001 Deadline Scheduling Results
27
Bob Givan Electrical and Computer Engineering Purdue University 26 November 4-9, 2001 Deadline Scheduling Results
28
Bob Givan Electrical and Computer Engineering Purdue University 27 November 4-9, 2001 Deadline Scheduling Results
29
Bob Givan Electrical and Computer Engineering Purdue University 28 November 4-9, 2001 Random Early Dropping Results Objective: minimize delay subject to throughput loss-tolerance Comparison: Candidate policies: RED and “buffer-k” KMN-sampling Rollout of buffer-k Parallel rollout Hindsight optimization Results due to H. S. Chang.
30
Bob Givan Electrical and Computer Engineering Purdue University 29 November 4-9, 2001 Random Early Dropping Results
31
Bob Givan Electrical and Computer Engineering Purdue University 30 November 4-9, 2001 Random Early Dropping Results
32
Bob Givan Electrical and Computer Engineering Purdue University 31 November 4-9, 2001 Congestion Control Results MDP Objective: minimize weighted sum of throughput, delay, and loss-rate Fairness is hard-wired Comparisons: PD-k (proportional-derivative with k target queue) Hindsight optimization Rollout of PD-k == parallel rollout Results due to G. Wu, in progress
33
Bob Givan Electrical and Computer Engineering Purdue University 32 November 4-9, 2001 Congestion Control Results
34
Bob Givan Electrical and Computer Engineering Purdue University 33 November 4-9, 2001 Congestion Control Results
35
Bob Givan Electrical and Computer Engineering Purdue University 34 November 4-9, 2001 Congestion Control Results
36
Bob Givan Electrical and Computer Engineering Purdue University 35 November 4-9, 2001 Congestion Control Results
37
Bob Givan Electrical and Computer Engineering Purdue University 36 November 4-9, 2001 Results Summary Unbiased sampling cannot cope Parallel rollout wins in 2 domains Not always equal to simple rollout of one base policy Hindsight optimization wins in 1 domain Simple policy rollout – the cheapest method Poor in domain 1 Strong in domain 2 with best base policy – but how to find this policy? So-so in domain 3 with any base policy
38
Bob Givan Electrical and Computer Engineering Purdue University 37 November 4-9, 2001 Talk Summary Case study of MDP sampling methods New methods offering practical improvements Parallel policy rollout Hindsight optimization Systematic methods for using traffic models to help make network control decisions Feasibility of real-time implementation depends on problem timescale
39
Bob Givan Electrical and Computer Engineering Purdue University 38 November 4-9, 2001 Ongoing Research Apply to other control problems (different timescales): Admission/access control QoS routing Link bandwidth allotment Multiclass connection management Problems arising in proxy-services Diagnosis and recovery
40
Bob Givan Electrical and Computer Engineering Purdue University 39 November 4-9, 2001 Ongoing Research (Cont’d) Alternative traffic models Multi-timescale models Long-range dependent models Closed-loop traffic Fluid models Learning traffic model online Adaptation to changing traffic conditions
41
Bob Givan Electrical and Computer Engineering Purdue University 40 November 4-9, 2001 Congestion Control (Cont’d)
42
Bob Givan Electrical and Computer Engineering Purdue University 41 November 4-9, 2001 Congestion Control Results
43
Bob Givan Electrical and Computer Engineering Purdue University 42 November 4-9, 2001 Hindsight Optimization (Cont’d)
44
Bob Givan Electrical and Computer Engineering Purdue University 43 November 4-9, 2001 Policy Rollout (Cont’d) Base Policy Policy-performanceperformance
45
Bob Givan Electrical and Computer Engineering Purdue University 44 November 4-9, 2001 Receding-horizon Control For large horizon H, policy is ~ stationary. At each time, if state is x, then apply action u * (x) = argmax a Q(x,a) = argmax a R(x,a) + E [V H-1 *(y)] Compute estimate of Q-value at each time.
46
Bob Givan Electrical and Computer Engineering Purdue University 45 November 4-9, 2001 Congestion Control (Cont’d)
47
Bob Givan Electrical and Computer Engineering Purdue University 46 November 4-9, 2001 Domain 3: Congestion Control High-priority traffic: Open-loop controlled Low-priority traffic: Closed-loop controlled Resources: Bandwidth and buffer Objective: optimize throughput, delay, loss, and fairness Bottleneck Node High-priority Traffic Best-effort Traffic...
48
Bob Givan Electrical and Computer Engineering Purdue University 47 November 4-9, 2001 Congestion Control Results
49
Bob Givan Electrical and Computer Engineering Purdue University 48 November 4-9, 2001 Congestion Control Results
50
Bob Givan Electrical and Computer Engineering Purdue University 49 November 4-9, 2001 Congestion Control Results
51
Bob Givan Electrical and Computer Engineering Purdue University 50 November 4-9, 2001 Congestion Control Results
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.