1 Solving Infinite Horizon Stochastic Optimization Problems John R. Birge Northwestern University (joint work with Chris Donohue, Xiaodong Xu, and Gongyun.

Slides:



Advertisements
Similar presentations
Optimal Adaptive Execution of Portfolio Transactions
Advertisements

Chp.4 Lifetime Portfolio Selection Under Uncertainty
Expected Utility and Post-Retirement Investment and Spending Strategies William F. Sharpe STANCO 25 Professor of Finance Stanford University
Risk Aversion and Capital Allocation to Risky Assets
L5: Dynamic Portfolio Management1 Lecture 5: Dynamic Portfolio Management The following topics will be covered: Will risks be washed out over time? Solve.
The securities market economy -- theory Abstracting again to the two- period analysis - - but to different states of payoff.
1 Dynamic portfolio optimization with stochastic programming TIØ4317, H2009.
Decision Theoretic Planning
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Nonlinear Programming
Separating Hyperplanes
P.V. VISWANATH FOR A FIRST COURSE IN INVESTMENTS.
Resource Management of Highly Configurable Tasks April 26, 2004 Jeffery P. HansenSourav Ghosh Raj RajkumarJohn P. Lehoczky Carnegie Mellon University.
Applications of Stochastic Programming in the Energy Industry Chonawee Supatgiat Research Group Enron Corp. INFORMS Houston Chapter Meeting August 2, 2001.
Behavioral Finance and Asset Pricing What effect does psychological bias (irrationality) have on asset demands and asset prices?
Infinite Horizon Problems
Planning under Uncertainty
SA-1 1 Probabilistic Robotics Planning and Control: Markov Decision Processes.
Investment. An Investor’s Perspective An investor has two choices in investment. Risk free asset and risky asset For simplicity, the return on risk free.
Markov Decision Processes
L9: Consumption, Saving, and Investments 1 Lecture 9: Consumption, Saving, and Investments The following topics will be covered: –Consumption and Saving.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
Multiagent Planning with Factored MDPs Carlos Guestrin Daphne Koller Stanford University Ronald Parr Duke University.
Department of Computer Science Undergraduate Events More
Chapter 13 Stochastic Optimal Control The state of the system is represented by a controlled stochastic process. Section 13.2 formulates a stochastic optimal.
Lecture 3: Arrow-Debreu Economy
Lecture 9 – Nonlinear Programming Models
Optimization of Linear Problems: Linear Programming (LP) © 2011 Daniel Kirschen and University of Washington 1.
Linear-Programming Applications
Linear Programming.
1 Capacity Planning under Uncertainty Capacity Planning under Uncertainty CHE: 5480 Economic Decision Making in the Process Industry Prof. Miguel Bagajewicz.
Instructor: Vincent Conitzer
Corporate Banking and Investment Risk tolerance and optimal portfolio choice Marek Musiela, BNP Paribas, London.
Generalized and Bounded Policy Iteration for Finitely Nested Interactive POMDPs: Scaling Up Ekhlas Sonu, Prashant Doshi Dept. of Computer Science University.
Optimal Gradual Annuitization: Quantifying the Cost of Switching to Annuities Optimal Gradual Annuitization: Quantifying the Cost of Switching to Annuities.
Managerial Decision Making and Problem Solving
7/2/2013Copyright R. Douglas Martin1 5. TRANSACTION COSTS AND MIP 5.1 Transaction costs constraints in MVO 5.2 Transaction costs penalty in quadratic utility.
Practical Dynamic Programming in Ljungqvist – Sargent (2004) Presented by Edson Silveira Sobrinho for Dynamic Macro class University of Houston Economics.
Reinforcement Learning 主講人:虞台文 Content Introduction Main Elements Markov Decision Process (MDP) Value Functions.
Approximate Dynamic Programming Methods for Resource Constrained Sensor Management John W. Fisher III, Jason L. Williams and Alan S. Willsky MIT CSAIL.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Nonlinear Programming Models
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Investment Performance Measurement, Risk Tolerance and Optimal Portfolio Choice Marek Musiela, BNP Paribas, London.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
Decision Theoretic Planning. Decisions Under Uncertainty  Some areas of AI (e.g., planning) focus on decision making in domains where the environment.
OR Chapter 8. General LP Problems Converting other forms to general LP problem : min c’x  - max (-c)’x   = by adding a nonnegative slack variable.
IT Applications for Decision Making. Operations Research Initiated in England during the world war II Make scientifically based decisions regarding the.
3 Components for a Spreadsheet Optimization Problem  There is one cell which can be identified as the Target or Set Cell, the single objective of the.
Optimal Investment and Consumption Strategies with Finite Horizon and Transaction Costs: Theory and Computation Min Dai Dept of Math National University.
CPS 570: Artificial Intelligence Markov decision processes, POMDPs
1 The economics of insurance demand and portfolio choice Lecture 1 Christian Gollier.
Department of Computer Science Undergraduate Events More
5.3 Mixed Integer Nonlinear Programming Models. A Typical MINLP Model.
DEPARTMENT/SEMESTER ME VII Sem COURSE NAME Operation Research Manav Rachna College of Engg.
Measuring Financial Investor’s Risk Aversion Guillaume Cousin, Be-Partner Philippe Delquié, INSEAD INFORMS San Antonio, November 2000.
Gambling as investment: a generalized Kelly criterion for optimal allocation of wealth among risky assets David C J McDonald Ming-Chien Sung Johnnie E.
1 2 Linear Programming Chapter 3 3 Chapter Objectives –Requirements for a linear programming model. –Graphical representation of linear models. –Linear.
Markov Decision Process (MDP)
Markov Decision Processes II
Theory of Capital Markets
Graphical Analysis – the Feasible Region
Markov Decision Processes
5.3 Mixed Integer Nonlinear Programming Models
Markov Decision Processes
William F. Sharpe STANCO 25 Professor of Finance
Mean-Variance Portfolio Rebalancing with Transaction Costs
Reinforcement Learning Dealing with Partial Observability
Dr. Arslan Ornek DETERMINISTIC OPTIMIZATION MODELS
Presentation transcript:

1 Solving Infinite Horizon Stochastic Optimization Problems John R. Birge Northwestern University (joint work with Chris Donohue, Xiaodong Xu, and Gongyun Zhao)

2 Motivating Problem (Very) long-term investor (example: university endowment) Payout from portfolio over time (want to keep payout from declining) Invest in various asset categories Decisions: How much to payout (consume)? How to invest in asset categories? Complication: restrictions on asset trades

3 Problem Formulation Notation: x – current state (x 2 X) u (or u x ) – current action given x (u (or u x ) 2 U(x))  – single period discount factor P x,u – probability measure on next period state y depending on x and u c(x,u) – objective value for current period given x and u V(x) – value function of optimal expected future rewards given current state x Problem: Find V such that V(x) = max u 2 U(x) {c(x,u) +  E Px,u [V(y)] } for all x 2 X.

4 Approach Define an upper bound on the value function V 0 (x) ¸ V(x) 8 x 2 X Iteration k: upper bound V k Solve for some x k TV k (x k ) = max u c(x k,u) +  E Pxk,u [V k (y)] Update to a better upper bound V k+1 Update uses an outer linear approximation on U k

5 Successive Outer Approximation V* V0V0 TV 0 x0x0 V1V1

6 Properties of Approximation V* · TV k · V k+1 · V k Contraction || TV k – V * || 1 ·  ||V k – V*|| 1 Unique Fixed Point TV*=V* ) if TV k ¸ V k, then V k =V*.

7 Convergence Value Iteration T k V 0 ! V* Distributed Value Iteration If you choose every x 2 X infinitely often, then V k ! V*. (Here, random choice of x, use concavity.) Deepest Cut Pick x k to maximize V k (x)-TV k (x) DC problem to solve Convergence again with continuity (caution on boundary of domain of V*)

8 Details for Random Choice Consider any x Choose i and x i s.t. ||x i – x|| <  i Suppose || r V i || · K 8 i || V k (x) – V*(x)|| · ||V k (x)-V k (x k )+V k (x k )- V*(x k )+V*(x k )-V*(x)|| · 2  k K +  ||V k-1 (x k )-V*(x k )|| · 2  i  i K +  k ||V 0 (x 0 ) – V*(x 0 )||

9 Cutting Plane Algorithm Initialization: Construct V 0 (x)=max u c 0 (x,u) +  E Px,u [V 0 (y)], where c 0 ¸ c and c 0 concave. V 0 is assumed piecewise linear and equivalent to V 0 (x)=max {  |  · E 0 x + e 0 }. k=0. Iteration: Sample x k 2 X (in any way such that the probability of x k 2 A is positive for any A ½ X of positive measure) and solve TV k (x k ) = max u c(x k,u) +  E Pxk,u [V k (y)] where V k (y) =max{  |  · E l y + e l,l=0,…,k.} Find supporting hyperplanes defined by E k+1 and e k+1 such that E k+1 x + e k+1 ¸ TV k (x). k à k+1. Repeat.

10 Specifying Algorithm Feasibility: Ax + Bu · b Transition: y=F i u for some realization i with probability p i Iteration k Problem: TV k (x k ) = max u,  c(x k,u) +   i p i  i s.t. A x k + B u · b, - E l (F i u) - e l +  i · 0, 8 i,l. From duality: TV k (x k ) = inf , l,i max u,  c(x k,u) -  (Ax k +Bu-b) +   i (p i  i +  l i,l (E l (F i u) + e l -  i )) · max u,  c(x k,u) -  k (Ax k +Bu-b) +   i (p i  i +  l i,l,k (E l (F i u) +e l -  i )) for optimal  k, i,l,k for x k · c(x k,u k ) + r c(x k,u k ) T (x-x k,-u k ) -  k Ax +  k b +  i (  l i,l,k e l ) Cuts: E k+1 = r x c(x k,u k ) T -  k A e k+1 equal to the constant terms.

11 Results Sometimes convergence is fast

12 Results Sometimes convergence is slow

13 Challenges Constrain feasible region to obtain convergence results Accelerate the DC search problem to find the deep cut Accelerate overall algorithm using: multiple simultaneous cuts? nonlinear cuts? bundles approach?

14 Conclusions Can formulate infinite-horizon investment problem in stochastic programming framework Solution with cutting plane method Convergence with some conditions Results for trade-restricted assets significantly different from market assets with same risk characteristics

15 Investment Problem Determine asset allocation and consumption policy to maximize the expected discounted utility of spending State and Action x=(cons, risky, wealth) u=(cons_new,risky_new) Two asset classes Risky asset, with lognormal return distribution Riskfree asset, with given return r f Power utility function Consumption rate constrained to be non-decreasing cons_new ¸ cons

16 Existing Research Dybvig ’95* Continuous-time approach Solution Analysis Consumption rate remains constant until wealth reaches a new maximum The risky asset allocation  is proportional to w-c/r f, which is the excess of wealth over the perpetuity value of current consumption  decreases as wealth decreases, approaching 0 as wealth approaches c/r f (which is in absence of risky investment sufficient to maintain consumption indefinitely). Dybvig ’01 Considered similar problem in which consumption rate can decrease but is penalized (soft constrained problem) * “Duesenberry's Ratcheting of Consumption: Optimal Dynamic Consumption and Investment Given Intolerance for any Decline in Standard of Living” Review of Economic Studies 62, 1995,

17 Objectives Replicate Dybvig continuous time results using discrete time approach Evaluate the effect of trading restrictions for certain asset classes (e.g., private equity) Consider additional problem features Transaction Costs Multiple risky assets

18 Results – Non-decreasing Consumption As number of time periods per year increases, solution converges to continuous time solution

19 Results – Non-Decreasing Consumption with Transaction Costs

20 Observations Effect of Trading Restrictions Continuously traded risky asset: 70% of portfolio for 4.2% payout rate Quarterly traded risky asset: 32% of portfolio for same payout rate Transaction Cost Effect Small differences in overall portfolio allocations Optimal mix depends on initial conditions

21 Extensions Soft constraint on decreasing consumption Allow some decreases with some penalty Lag on sales Waiting period on sale of risky assets (e.g., 60- day period) Multiple assets Allocation bounds

22 Approach Application of typical stochastic programming approach complicated by infinite horizon Initialization. Define a valid constraint on Q(x) Requires problem knowledge. For optimal consumption problem, assume extremely high rate of consumption forever

23 Approach (cont.) Iteration k Expensive search over x, possible for the optimal consumption problem because of small number of variables