1 A Stochastic Differential (Difference) Game Model With an LP Subroutine for Mixed and Pure Strategy Optimization INFORMS International Meeting 2007,

Slides:



Advertisements
Similar presentations
Chapter 9 Maintenance and Replacement The problem of determining the lifetime of an asset or an activity simultaneously with its management during that.
Advertisements

The simplex algorithm The simplex algorithm is the classical method for solving linear programs. Its running time is not polynomial in the worst case.
9.1 Strictly Determined Games Game theory is a relatively new branch of mathematics designed to help people who are in conflict situations determine the.
GAME THEORY.
Linear Programming Problem. Introduction Linear Programming was developed by George B Dantzing in 1947 for solving military logistic operations.
1 1 Slide © 2009 South-Western, a part of Cengage Learning Slides by John Loucks St. Edward’s University.
Two-Player Zero-Sum Games
Operations Research Assistant Professor Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD4207 University of Palestine.
Part 3: The Minimax Theorem
Boyce/DiPrima 10th ed, Ch 10.1: Two-Point Boundary Value Problems Elementary Differential Equations and Boundary Value Problems, 10th edition, by William.
MDP Presentation CS594 Automated Optimal Decision Making Sohail M Yousof Advanced Artificial Intelligence.
Planning under Uncertainty
POMDPs: Partially Observable Markov Decision Processes Advanced AI
Tools for optimal coordination of CCS, power industry capacity expansion and bio energy raw material production and harvesting Peter Lohmander Prof. Dr.
Using process knowledge to identify uncontrolled variables and control variables as inputs for Process Improvement 1.
OPTIMAL PRESENT RESOURCE EXTRACTION UNDER THE INFLUENCE OF FUTURE RISK Professor Dr Peter Lohmander SLU, Sweden,
Chapter 4 Multiple Regression.
Extensions to Consumer theory Inter-temporal choice Uncertainty Revealed preferences.
An Introduction to Game Theory Part III: Strictly Competitive Games Bernhard Nebel.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 68 Chapter 9 The Theory of Games.
Games in the normal form- An application: “An Economic Theory of Democracy” Carl Henrik Knutsen 5/
Lecture Slides Elementary Statistics Twelfth Edition
Nonlinear Filtering of Stochastic Volatility Aleksandar Zatezalo, University of Rijeka, Faculty of Philosophy in Rijeka, Rijeka, Croatia
Game Theory Objectives:
Chapter 13 Stochastic Optimal Control The state of the system is represented by a controlled stochastic process. Section 13.2 formulates a stochastic optimal.
Concession Length and Investment Timing Flexibility Chiara D’Alpaos, Cesare Dosi and Michele Moretto.
Engineering Systems Analysis for Design Richard de Neufville © Massachusetts Institute of Technology Flaw of Averages Slide 1 of 29 Richard de Neufville.
Game Theory.
CS Reinforcement Learning1 Reinforcement Learning Variation on Supervised Learning Exact target outputs are not given Some variation of reward is.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
MAKING COMPLEX DEClSlONS
Chapter 12 & Module E Decision Theory & Game Theory.
Auction Seminar Optimal Mechanism Presentation by: Alon Resler Supervised by: Amos Fiat.
Optimal continuous cover forest management: - Economic and environmental effects and legal considerations Professor Dr Peter Lohmander
Management of Waiting Lines McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
1 CM Optimal Resource Control Model & General continuous time optimal control model of a forest resource, comparative dynamics and CO2 storage consideration.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 49 DP can give complete quantitative solution Example 1: Discrete, finite.
1 Optimal Forest Management Under Risk International Seminars in Life Sciences Universidad Politécnica de Valencia Thursday Peter Lohmander.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
In section 11.9, we were able to find power series representations for a certain restricted class of functions. Here, we investigate more general problems.
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
MA4248 Weeks 6-7. Topics Work and Energy, Single Particle Constraints, Multiple Particle Constraints, The Principle of Virtual Work and d’Alembert’s Principle.
1 Optimal activities over time in a typical forest industry company in the presence of stochastic markets - A flexible approach and possible extensions!
1. 2 Some details on the Simplex Method approach 2x2 games 2xn and mx2 games Recall: First try pure strategies. If there are no saddle points use mixed.
Optimal continuous natural resource extraction with increasing risk in prices and stock dynamics Professor Dr Peter Lohmander
Copyright © Cengage Learning. All rights reserved. 9 Inferences Based on Two Samples.
Markov Decision Process (MDP)
Topics 1 Specific topics to be covered are: Discrete-time signals Z-transforms Sampling and reconstruction Aliasing and anti-aliasing filters Sampled-data.
Management of Waiting Lines Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent.
2.5 The Fundamental Theorem of Game Theory For any 2-person zero-sum game there exists a pair (x*,y*) in S  T such that min {x*V. j : j=1,...,n} =
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Optimization of adaptive control functions in multidimensional forest management via stochastic simulation and grid search Version Peter Lohmander.
Keynote presentation at:
Stochastic dynamic programming with Markov chains
Market-Risk Measurement
Lecture Slides Elementary Statistics Eleventh Edition
Chapter 15: Game Theory: The Mathematics Lesson Plan of Competition
Tools for Decision Analysis: Analysis of Risky Decisions
Ch 10.1: Two-Point Boundary Value Problems
Goodness of Fit x² -Test
Hidden Markov Models Part 2: Algorithms
Lohmander, P., Adaptive Optimization of Forest
Lecture Slides Elementary Statistics Twelfth Edition
Lecture Slides Elementary Statistics Twelfth Edition
Optimal stochastic control in continuous time with Wiener processes: - General results and applications to optimal wildlife management KEYNOTE Version.
Laplace Transforms Lecture-11 Additional chapters of mathematics
Chapter 15: Game Theory: The Mathematics Lesson Plan of Competition
Lecture Slides Essentials of Statistics 5th Edition
Lecture Slides Essentials of Statistics 5th Edition
Presentation transcript:

1 A Stochastic Differential (Difference) Game Model With an LP Subroutine for Mixed and Pure Strategy Optimization INFORMS International Meeting 2007, Puerto Rico Peter Lohmander Professor SLU Umea, SE , Sweden, Version

2 Abstract: This paper presents a stochastic two person differential (difference) game model with a linear programming subroutine that is used to optimize pure and/or mixed strategies as a function of state and time. In ”classical dynamic models”, ”pure strategies” are often assumed to be optimal and deterministic continuous path equations can be derived. In such models, scale and timing effects in operations are however usually not considered. When ”strictly convex scale effects” in operations with defence and attack (”counter air” or ”ground support”) are taken into consideration, dynamic mixed OR pure strategies are optimal in different states. The optimal decision frequences are also functions of the relative importance of the results from operations in different periods. The expected cost of forcing one participant in the conflict to use only pure strategies is determined. The optimal responses of the opposition in case one participant in the conflict is forced to use only pure strategies are calculated. Dynamic models of the presented type, that include mixed strategies, are highly relevant and must be used to determine the optimal strategies. Otherwise, considerable and quite unnecessary losses should be expected.

3 Inspiration: Isaacs R. (1965) Differential Games – A mathematical theory with applications to warfare and pursuit, control and optimization, Wiley, 1965 (Also Dover, 1999) Washburn, A.R. (2003) Two-person zero-sum games, 3 ed., INFORMS, Military applications society, Topics in operations research

4 Flexibility Some resources may be used in different ways. This is true and important in the air force, in the army and in the enterprises in the economy.

5 Our Plan to Win This plan is focused on continuously increasing McDonald's relevance to consumers' everyday lives through multiple initiatives surrounding the five key drivers of great restaurant experiences. Our efforts, complemented with financial discipline, have delivered for shareholders some of the strongest results in our history.

6 W-lan Max Hamburger restaurants is the only national restaurant chain to offer free Internet access for all guests. Connection is offered via W- Lan. The most satisfied guests For the fith year in a row Max Hamburger Restaurants has Sweden's most satisfied customers among the national hamburger restaurants according to a new survey by SIFO/ISI Wissing.

7 Boeing B-52 Stratofortress Intercontinental Strategic Bomber

8 JAS 39 Gripen Multirole fighter

9

10

11

12

13

14 Special case 1: Special case 2:

15 In Model 1:

16 In Model 2:

17 In Model 3 and Model 4:

18 Model 1

19

20 Model 1 !TAW_070611; !Peter Lohmander; Model: sets: xset/1..3/:; yset/1..3/:; nset/1..3/:; D1set/1..3/:; D2set/1..3/:; xyset(xset,yset):; xynset(xset,yset,nset):V; xyD1set(xset,yset,D1set):GS1, CA1; xyD2set(xset,yset,D2set):GS2, CA2; xynD1s(xset,yset,nset,D1set):xdec; xD1D2s(xset, D1set, D2set):; xD1D2Xs(xset,D1set, D2set, xset):TPX; yD2D1s(yset,D2set,D1set):; yD2D1yset(yset,D2set,D1set,yset):TPY; endsets

21 max = Z; Z !Terminal condition: No more activities occur when n <= 1. = 0);

22 ! The last day when activities occur, n=2, all resources should be used for GS. = x-y);

23 ! Day n >= 3, you may use some resources for GS and some for CA. The rows and columns represent the resources used for GS. #LE# x: GS1(x,y,D1) = D1-1 #GT# x: GS1(x,y,D1) = 0 CA1(x,y,D1) = x GS1(x,y,D1) #LE# y: GS2(x,y,D2) = D2-1 #GT# y: GS2(x,y,D2) = 0 CA2(x,y,D2) = y GS2(x,y,D2));

24 = D1#GT#x: xdec(x,y,n,D1) = 0);

25 !Sum (Transition probability) TPX(x1,GS1,CA2,x2) ) = TPY(y1,GS2,CA1,y2) ) = 1 );

26 !********************************************; ! Transition probability calculations for x ; !********************************************; ! Transition probabilities for x if x1 = : TPX(x1,GS1P1,CA2P1,x1) = 1 ); ! Transition probabilities for x if CA2 = #EQ# 1 : TPX(x1,GS1P1,CA2P1,x1) = 1 ); ! Transition probabilities for x if GS1 = #EQ# 1 : TPX(x1,GS1P1,CA2P1,x1) = 1 );

27 ! Transition probabilities for x if GS1 >= 1 and CA2 = #AND# GS1P1#GE#2 #AND# CA2P1#EQ#2 : TPX(x1,GS1P1,CA2P1,x1) = 1/2 #AND# GS1P1#GE#2 #AND# CA2P1#EQ#2 : TPX(x1,GS1P1,CA2P1,x1-1) = 1/2 );

28 ! Transition probabilities for x if GS1 = 1 and CA2 = #AND# GS1P1#EQ#2 #AND# CA2P1 #EQ# 3 : TPX(x1,GS1P1,CA2P1,x1) = 1/4 #AND# GS1P1#EQ#2 #AND# CA2P1 #EQ# 3 : TPX(x1,GS1P1,CA2P1,x1-1) = 3/4 );

29 ! Transition probabilities for x if GS1 >= 2 and CA2 = #AND# GS1P1#GE#3 #AND# CA2P1 #EQ# 3 : TPX(x1,GS1P1,CA2P1,x1) = 1/4 #AND# GS1P1#GE#3 #AND# CA2P1 #EQ# 3 : TPX(x1,GS1P1,CA2P1,x1-1) = 1/2 #AND# GS1P1#GE#3 #AND# CA2P1 #EQ# 3 : TPX(x1,GS1P1,CA2P1,x1-2) = 1/4 );

30 !*******************************************; ! Transition probability calculations for y ; !*******************************************; ! Transition probabilities for y if y1 = : TPY(y1,GS2P1,CA1P1,y1) = 1 ); ! Transition probabilities for y if CA1 = #EQ# 1 : TPY(y1,GS2P1,CA1P1,y1) = 1 ); ! Transition probabilities for y if GS2 = #EQ# 1 : TPY(y1,GS2P1,CA1P1,y1) = 1 );

31 ! Transition probabilities for y if GS2 >= 1 and CA1 = #AND# GS2P1#GE#2 #AND# CA1P1#EQ#2 : TPY(y1,GS2P1,CA1P1,y1) = 1/2 #AND# GS2P1#GE#2 #AND# CA1P1#EQ#2 : TPY(y1,GS2P1,CA1P1,y1-1) = 1/2 );

32 ! Transition probabilities for y if GS2 = 1 and CA1 = #AND# GS2P1#EQ#2 #AND# CA1P1#EQ#3 : TPY(y1,GS2P1,CA1P1,y1) = 1/4 #AND# GS2P1#EQ#2 #AND# CA1P1#EQ#3 : TPY(y1,GS2P1,CA1P1,y1-1) = 3/4 );

33 ! Transition probabilities for y if GS2 >= 2 and CA1 = #AND# GS2P1#GE#3 #AND# CA1P1 #EQ# 3 : TPY(y1,GS2P1,CA1P1,y1) = 1/4 #AND# GS2P1#GE#3 #AND# CA1P1 #EQ# 3 : TPY(y1,GS2P1,CA1P1,y1-1) = 1/2 #AND# GS2P1#GE#3 #AND# CA1P1 #EQ# 3 : TPY(y1,GS2P1,CA1P1,y1-2) = 1/4 );

34 In Model 1:

35 In Model 3 and Model 4:

36 !**********************; ! The game problems #LE# y: [ydec] V(x,y,3) (GS1(x,y,D1)/(1+CA2(x,y,D2)) - GS2(x,y,D2)/(1+CA1(x,y,D1)) TPX(x,GS1(x,y,D1)+1,CA2(x,y,D2)+1,x2)*TPY(y,GS2(x,y,D2)+1,CA1(x,y,D1)+1,y2)* V(x2,y2,2) )) * xdec(x,y,3,D1) ) );

37 end

38 In Model 3 and Model 4:

39 Model 3; (x,y,t) = (2,2,T-1) GS2 = 0 CA2 = 2 GS2 = 1 CA2 = 1 GS2 = 2 CA2 = 0 GS1 = 0 CA1 = 2 3/2*(0-0) + 0 = 0 3/2*(0 – (5/3)/(1+4)) *(0) *(+1) = /2*(0 – (5/3)*2/(1+4)) *(0) + 0.5*(1) *(2) = 0 GS1 = 1 CA1 = 1 3/2*(1/(1+4)-0) *(0) *(-1) = /2*(1/(1+1) – (5/3)/(1+1)) + 0.5*(0)+0.25*(1) *(-1) = /2*(1/(1+0) – (5/3)*2/(1+1)) + 0.5*(0) + 0.5*(1) = -0.5 GS1 = 2 CA1 = 0 3/2*(4/(1+4) – 0) *(0) + 0.5*(-1) *(-2) = 0.2 3/2*(4/(1+1) – (5/3)/(1+0)) + 0.5*(0) + 0.5*(-1) = 0 3/2*(4/(1+0)-(5/3)*2/(1+0)) + 0 = 1

40 Model 3; (x,y,t) = (2,2,T-1) GS2 = 0 CA2 = 2 GS2 = 1 CA2 = 1 GS2 = 2 CA2 = 0 GS1 = 0 CA1 = GS1 = 1 CA1 = GS1 = 2 CA1 =

41

42

43

44

45 Results from Model 1

46 Table 1.1 V(x,y,3) with Model 1 (This (n=3) means that t =T-1 since we use backward recursion and period T+1 is defined as n=1. )

47

48 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(1,1)(1,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0) Table 1.2 (GS1*, GS2*) at n=3 with Model 1

49 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(1,1)(1,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0) Table 1.2 (GS1*, GS2*) at n=3 with Model 1

50

51 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(1,1)(1,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0) Table 1.2 (GS1*, GS2*) at n=3 with Model 1

52

53 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(1,1)(1,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0) Table 1.2 (GS1*, GS2*) at n=3 with Model 1

54

55 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(1,1)(1,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0) Table 1.2 (GS1*, GS2*) at n=3 with Model 1

56

57 Model 2

58 In Model 2:

59 Model 2 !**********************; ! The game problems #LE# y: [ydec] V(x,y,3) (GS1(x,y,D1)*GS1(x,y,D1)/(1+CA2(x,y,D2)*CA2(x,y,D2)) - 5/3*GS2(x,y,D2)/(1+CA1(x,y,D1)*CA1(x,y,D1)) TPX(x,GS1(x,y,D1)+1,CA2(x,y,D2)+1,x2)*TPY(y,GS2(x,y,D2)+1,CA1(x,y,D1)+1,y2)* V(x2,y2,2) )) * xdec(x,y,3,D1) ) );

60 Results from Model 2

61 Table 2.1 V(x,y,3) with Model

62 Table 2.2 (GS1*, GS2*) at n=3 with Model 2 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

63 Table 2.2 (GS1*, GS2*) at n=3 with Model 2 (0,2)(0,2)(0,2)(0,2)(0,0)(0,0) (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

64

65 Model 3

66 In Model 3 and Model 4:

67 !**********************; ! The game problems #LE# y: [ydec] V(x,y,3) ( 3/2*(GS1(x,y,D1)*GS1(x,y,D1)/(1+CA2(x,y,D2)*CA2(x,y,D2)) - 5/3*GS2(x,y,D2)/(1+CA1(x,y,D1)*CA1(x,y,D1)) ) !A constant is added and later removed; + (1 TPX(x,GS1(x,y,D1)+1,CA2(x,y,D2)+1,x2)*TPY(y,GS2(x,y,D2)+1,CA1(x,y,D1)+1,y2)* V(x2,y2,2) )) * xdec(x,y,3,D1) ) );

68 Results from Model 3

69 Table 3.1 V(x,y,3) with Model

70 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

71 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

72 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

73

74 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

75

76 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

77

78 Table 3.2 (GS1*, GS2*) at n=3 with Model 3 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

79

80 Model 4

81 In Model 3 and Model 4:

82 !**********************; ! The game problems #LE# y: [ydec] V(x,y,3) ( 3/2*(GS1(x,y,D1)*GS1(x,y,D1)/(1+CA2(x,y,D2)*CA2(x,y,D2)) - 5/3*GS2(x,y,D2)/(1+CA1(x,y,D1)*CA1(x,y,D1)) ) TPX(x,GS1(x,y,D1)+1,CA2(x,y,D2)+1,x2)*TPY(y,GS2(x,y,D2)+1,CA1(x,y,D1)+1,y2)* V(x2,y2,2) )) * xdec(x,y,3,D1) ) ); Note that “discounting” has been introduced!

83 Results from Model 4

84 Table 4.1 V(x,y,3) with Model

85 Table 4.2 (GS1*, GS2*) at n=3 with Model 4 (0,2)(0,2)(0,2)(0,2) (0,0) Prob = * = (2,0) Prob = * = (0,1) Prob = * = (2,1) Prob = * = Sum = (0,1)(0,1)(0,1)(0,1)(2,0)(2,0) (0,0)(0,0)(1,0)(1,0)(2,0)(2,0)

86 Detailed study of optimal strategies: Optimal strategies with Model 3 (In Model 3, future results are not discounted. d = 1) P1: StrategyProbability Full CA Full GS P2: StrategyProbability Full CA % CA & 50% GS

87 Detailed study of optimal strategies: Optimal strategies with Model 4 (In Model 3, future results are not discounted. d = 0.85) P1: P2: StrategyProbability Full CA2/3 Full GS1/3 StrategyProbability Full CA % CA & 50% GS

88 When the importance of instant results in relation to future results increases (= when the discounting of future results increases from 0 to 15%): P1 increases the probability of instant full CA. P1 decreases the probability of instant GS. P2 decreases the probability of instant full CA. P2 increases the probability of partial CA and partial GS.

89 Observations: Even in case there is just one more period of conflict ahead of us and even if the participants only have two units available per participant, we find that the optimal strategies are mixed. The optimal decision frequencies are affected by the result discounting. The different participants should optimally change the strategies in qualitatively different ways when the degree of discounting changes.

90 Observations cont.: Differential games in continuous time can not describe these mixed strategies. Furthermore, even if we would replace deterministic differential games in continuous time by “stochastic differential games” based on stochastic differential equations in continuous time, this would not capture the real situation, since the optimal frequencies are results of the scales of missions. Two resource units used during one time interval will not give the same result as one resource unit during two time intervals.

91 Isaacs (1965), in section 5.4, The war of attrition and attack: “The realistic execution of the strategies would comprise of a series of discrete decisions. But we shall smooth matters into a continuous process. Such is certainly no farther from truth than our assumptions and therefore can be expected to be as reliable as a stepped model. It is also more facile to handle and yields general results more readily.”

92 Isaacs (1965), page 66: ”Similarily, except possibly on singular surfaces, the optimal strategies will be continuous functions denoted by... when they are unique; when they are not, we shall often assume such continuous functions can be selected.” Observation (Lohmander): With economies of scale in operations, the optimal strategies will usually not be continuous functions!

93 V u X=2 X=1 X=0

94 V* x

95 u*(x) x

96 Isaacs (1965) obtains differential equations that describe the optimal decisions over time by the two participants: The system moves according to these equations:

97 The objective function in Isaacs (1965) is linear in the decision variables and time is continuous:

98 There is no reason to expect that the Isaacs (1965) model would lead to mixed strategies. The objective function is linear. Furthermore there are no scale effects in the “dynamic transition”.

99 Washburh (2003), Tactical air war: ”- It is a curious fact that in no case is a mixed strategy ever needed; in fact, in all cases each player assigns either all or none of his air force to GS.” (The model analysed by Washburn has very strong similarities to the model analysed here. The Washburn model was even the starting point in the development of this model. However, the objective function used by Washburn is linear. Hence, the results become quite different.)

100 An analogy to optimal control: The following observations are typical when optimal management of natural resources is investigated: If the product price is a constant or a decreasing function of the production and sales volume and if the cost of extraction is a strictly convex function of the extraction volume, the objective function is strictly concave in the production volume. Then we obtain a smooth differentiable optimal time path of the resource stock. However, if the extraction cost is concave, which is sometimes the case if you have considerable set up costs or other scale advantages, then the optimal strategy is usually “pulse harvesting”. In this case, the objective function that we are maximizing is convex in the decision variable. The optimal stock level path then becomes non differentiable.

101 Technology differences: The reason why the two participants should make different decisions with different strategies is that they have different “GS-technologies”. These technologies have different degree of convexity of results with respect to the amount of resources used. Such differences are typical in real conflicts. They are caused by differences in equipment and applied methods. Of course, in most cases we can expect that the applied methods are adapted to the equipment and other relevant conditions.

102 For P1, the instant GS result is strictly convex in the number of units devoted to the GS mission. For P2, the instant GS result is proportional to the number of units devoted to the GS mission.

103 Model 5

104 Model 5 corresponds to Model 3 but the following line is added just before “end”. Case 1: xdec(3,3,3,1) = 1; Case 2: xdec(3,3,3,2) = 1; Case 3: xdec(3,3,3,3) = 1;

105 Results from Model 5

106 Results if the decisions of P1 are not constrained in (2,2) when n = 3 (Model 3): Case 0: xdec(3,3,3,1) = xdec(3,3,3,3) = ; (GS1 = 0 and CA1 = 2 or GS1 = 2 and CA1 = 0 ) Effects: The value of the game : V(3,3) = The opposition should use a mixed strategy: GS2 = 0, CA2 = 2 with probability GS2=1, CA2 = 1 with probability

107 Results if the decisions of P1 are constrained in (2,2) when n = 3: Case 1: xdec(3,3,3,1) = 1; (GS1 = 0 and CA1 = 2 ) Effects: The value of the game decreases: V(3,3) = 0 The decision of the opposition changes: GS2 = 2, CA2 = 0

108 Results if the decisions of P1 are constrained in (2,2) when n = 3: Case 2: xdec(3,3,3,2) = 1; (GS1 = 1 and CA1 =1 ) Effects: The value of the game decreases: V(3,3) = -0.5 The decision of the opposition changes: GS2 = 2, CA2 = 0

109 Results if the decisions of P1 are constrained in (2,2) when n = 3: Case 3: xdec(3,3,3,3) = 1; (GS1 = 2 and CA1 = 0 ) Effects: The value of the game decreases: V(3,3) = 0 The decision of the opposition changes: GS2 = 1, CA2 = 1

110 Conclusions from Model 5: Since we are interested in the best attainable expected outcome, we should not restrict our decisions to a pure strategy and let the opposition select strategy without such restrictions! You should avoid being predictable!

111 Conclusions: This paper presents a stochastic two person differential (difference) game model with a linear programming subroutine that is used to optimize pure and/or mixed strategies as a function of state and time. In ”classical dynamic models”, ”pure strategies” are often assumed to be optimal and deterministic continuous path equations can be derived. In such models, scale and timing effects in operations are however usually not considered. When ”strictly convex scale effects” in operations with defence and attack (”counter air” or ”ground support”) are taken into consideration, dynamic mixed OR pure strategies are optimal in different states. The optimal decision frequences are also functions of the relative importance of the results from operations in different periods. The expected cost of forcing one participant in the conflict to use only pure strategies is determined. The optimal responses of the opposition in case one participant in the conflict is forced to use only pure strategies are calculated. Dynamic models of the presented type, that include mixed strategies, are highly relevant and must be used to determine the optimal strategies. Otherwise, considerable and quite unnecessary losses should be expected.

112 Contact: Peter Lohmander Professor SLU SE Umea, Sweden Home page:

113 Source: Lohmander, P., TAW models doc