APEC 8205: Applied Game Theory Fall 2007

Slides:



Advertisements
Similar presentations
Vincent Conitzer CPS Repeated games Vincent Conitzer
Advertisements

Some Problems from Chapt 13
Infinitely Repeated Games
M9302 Mathematical Models in Economics Instructor: Georgi Burlakov 3.1.Dynamic Games of Complete but Imperfect Information Lecture
Crime, Punishment, and Forgiveness
© 2009 Institute of Information Management National Chiao Tung University Game theory The study of multiperson decisions Four types of games Static games.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Game Theory “Доверяй, Но Проверяй” - Russian Proverb (Trust, but Verify) - Ronald Reagan Mike Shor Lecture 6.
Oligopoly Games An Oligopoly Price-Fixing Game
Game Theory “Доверяй, Но Проверяй” (“Trust, but Verify”) - Russian Proverb (Ronald Reagan) Topic 5 Repeated Games.
Chapter Twenty-Eight Game Theory. u Game theory models strategic behavior by agents who understand that their actions affect the actions of other agents.
Static Games and Cournot Competition
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
Non-Cooperative Game Theory To define a game, you need to know three things: –The set of players –The strategy sets of the players (i.e., the actions they.
Chapter 14 Infinite Horizon 1.Markov Games 2.Markov Solutions 3.Infinite Horizon Repeated Games 4.Trigger Strategy Solutions 5.Investing in Strategic Capital.
The basics of Game Theory Understanding strategic behaviour.
M9302 Mathematical Models in Economics Instructor: Georgi Burlakov 2.5.Repeated Games Lecture
Infinitely Repeated Games Econ 171. Finitely Repeated Game Take any game play it, then play it again, for a specified number of times. The game that is.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
Game Theory: Inside Oligopoly
Game Theory Lecture 9.
Game Theory Lecture 8.
Repeated Prisoner’s Dilemma If the Prisoner’s Dilemma is repeated, cooperation can come from strategies including: “Grim Trigger” Strategy – one.
Games People Play. 8: The Prisoners’ Dilemma and repeated games In this section we shall learn How repeated play of a game opens up many new strategic.
Dynamic Games of Complete Information.. Repeated games Best understood class of dynamic games Past play cannot influence feasible actions or payoff functions.
Objectives © Pearson Education, 2005 Oligopoly LUBS1940: Topic 7.
Yale Lectures 21 and Repeated Games: Cooperation vs the End Game.
Final Lecture. ``Life can only be understood backwards; but it must be lived forwards.” Søren Kierkegaard Thoughts on subgame perfection?
Chapter Twenty-Eight Game Theory. u Game theory models strategic behavior by agents who understand that their actions affect the actions of other agents.
Static Games of Complete Information: Equilibrium Concepts
Static Games and Cournot Competition
Repeated games - example This stage game is played 2 times Any SPNE where players behave differently than in a 1-time game? Player 2 LMR L1, 10, 05, 0.
TOPIC 6 REPEATED GAMES The same players play the same game G period after period. Before playing in one period they perfectly observe the actions chosen.
Static Games of Complete Information: Subgame Perfection
Lecture 3: Oligopoly and Strategic Behavior Few Firms in the Market: Each aware of others’ actions Each firm in the industry has market power Entry is.
0 MBA 299 – Section Notes 4/25/03 Haas School of Business, UC Berkeley Rawley.
Two-Stage Games APEC 8205: Applied Game Theory Fall 2007.
© 2009 Institute of Information Management National Chiao Tung University Lecture Notes II-2 Dynamic Games of Complete Information Extensive Form Representation.
Problems from Chapter 12. Problem 1, Chapter 12 Find a separating equilibrium Trial-and-error. Two possible separating strategies for Player 1: – Choose.
1 Game Theory Sequential bargaining and Repeated Games Univ. Prof.dr. M.C.W. Janssen University of Vienna Winter semester Week 46 (November 14-15)
Punishment and Forgiveness in Repeated Games. A review of present values.
Nash equilibrium Nash equilibrium is defined in terms of strategies, not payoffs Every player is best responding simultaneously (everyone optimizes) This.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Chapter 9: Static Games and Cournot Competition 1 Static Games and Cournot Competition.
Dynamic Games & The Extensive Form
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Topic 3 Games in Extensive Form 1. A. Perfect Information Games in Extensive Form. 1 RaiseFold Raise (0,0) (-1,1) Raise (1,-1) (-1,1)(2,-2) 2.
Lecture 5 Leadership and Reputation Reputations arise in situations where there is an element of repetition, and also where coordination between players.
제 10 장 게임이론 Game Theory: Inside Oligopoly
Punishment, Detection, and Forgiveness in Repeated Games.
Final Lecture. Problem 2, Chapter 13 Exploring the problem Note that c, x yields the highest total payoff of 7 for each player. Is this a Nash equilibrium?
Mixed Strategies and Repeated Games
Oligopoly Theory1 Oligopoly Theory (11) Collusion Aim of this lecture (1) To understand the idea of repeated game. (2) To understand the idea of the stability.
Game Theory (Microeconomic Theory (IV)) Instructor: Yongqin Wang School of Economics, Fudan University December, 2004.
Dynamic games, Stackelburg Cournot and Bertrand
Topics to be Discussed Gaming and Strategic Decisions
Lec 23 Chapter 28 Game Theory.
Entry Deterrence Players Two firms, entrant and incumbent Order of play Entrant decides to enter or stay out. If entrant enters, incumbent decides to fight.
1 Strategic Thinking Lecture 7: Repeated Strategic Situations Suggested reading: Dixit and Skeath, ch. 9 University of East Anglia School of Economics.
ECON 330 Lecture 17 Monday, November 25.
Dynamic Games of Complete Information
Vincent Conitzer CPS Repeated games Vincent Conitzer
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Multiagent Systems Repeated Games © Manfred Huber 2018.
Vincent Conitzer Repeated games Vincent Conitzer
Chapter 14 & 15 Repeated Games.
Chapter 14 & 15 Repeated Games.
Molly W. Dahl Georgetown University Econ 101 – Spring 2009
Lecture Game Theory.
Vincent Conitzer CPS Repeated games Vincent Conitzer
Presentation transcript:

APEC 8205: Applied Game Theory Fall 2007 Repeated Games APEC 8205: Applied Game Theory Fall 2007

Objectives Understand the Class of Repeated Games Understand Conditions Under Which Non-Nash Play Can be Sustained as a Subgame Perfect Nash Equilibrium when a Game is Repeated Multiple Nash Equilibria Infinite Repetition

Why study repeated games? Many interactions in life are repeated. Large retailers compete on a daily basis for customers. Dana and I compete on a daily basis to decide who will make dinner and who will pickup around the house. Mason and Spencer compete on a daily basis to see who gets to watch TV and who gets to play X-Box. What is of interest in these type of repeated interactions? Can players achieve better results than might occur in a single shot game? Can players use the history of play to their advantage?

Some Terminology G: Stage game (usually thought of in normal form). Players: i = 1,..,N ai  Ai: Strategy space for player i. a = (a1,…,aN)  A = i = 1NAi: Strategy profile for all players. ui(a): Player i’s payoff for strategy profile a. u(a) = (u1(a),…, uN(a)): Vector of player payoffs for strategy profile a. T: Number of times the stage game is repeated (could be infinite). ait  Ai : Player i’s strategy choice at time t. at = (a1t,…,aNt)  A = i = 1NAi: Strategy profile for all players at time t. ht = (a1,…,at-1)  At = t’ = 1t-1A: History of play at time t. sit(ht)  Ai: History dependent strategy. st(ht) = (s1t(ht), …, sNt(ht))  A: History dependent strategy profile. Ui(s1(h1),…, sT(hT)) = t=1Twitui(st(ht)): Player i’s payoff from the game. U(s1(h1),…, sT(hT)) = (U1(s1(h1),…, sT(hT)),…,UN(s1(h1),…, sT(hT))): Payoffs for all players.

Consider and Example Suppose this Prisoner’s Dilemma game is played twice and that wit =1 for i = 1,2 and t = 1,2.

Two Period Prisoner’s Dilemma Example In Extensive Form Player 1 C D Player 2 C D C D Player 1 Player 1 Player 1 Player 1 C D C D C D C D Player 2 Player 2 Player 2 Player 2 C D C D C D C D C D C D C D C D 4 2 5 5 2 3 2 5 6 3 1 4 5 2 3 6 4 1 3 1 4 4 1 2 Player 1’s Payoff Player 2’s Payoff

Two Period Prisoner’s Dilemma Example After Solving Stage 2 Subgames Player 1 C D Player 2 C D C D Player 1 Player 1 Player 1 Player 1 D D D D Player 2 Player 2 Player 2 Player 2 D D D D 3 1 4 4 1 2 Player 1’s Payoff Player 2’s Payoff

Two Period Prisoner’s Dilemma Example After Solving Game As Whole Player 1 D Player 2 Therefore, the subgame perfect strategies are (strategy choice in stage 1, strategy choice in stage 2 given (D,D) in stage 1, strategy choice in stage 2 given (D,C) in stage 1, strategy choice in stage 2 given (C,D) in stage 1, strategy choice in stage 2 given (C,C) in stage 1) = (D,D,D,D,D) for both players. D Player 1 D Player 2 D 2 Player 1’s Payoff Player 2’s Payoff

So, what is the point? If the stage game of a finitely repeated game has a unique Nash equilibrium, then there is a unique subgame perfect equilibrium where that Nash equilibrium is played in every stage of the game! But what can happen if there is not a unique equilibrium? Or what if the stage game can be infinitely repeated?

What about multiple equilibria? Consider this modified version of the Prisoner’s Dilemma and assume T = 2 and wit = 1 for i = 1,2 and t = 1,2.

Starting with Period 2 There are 9 possible histories for the 2nd period of this game: (U,L), (U,C), (U,R), (M,L), (M,C), (M,R), (D,L), (D,C), and (D,R). For any subgame starting from one of these histories, there are two potential Nash equilibria: (M,C) or (D,R). Therefore, for an equilibrium strategy to be subgame perfect, it must have (M,C) or (R,D) in response to the history (x, y) for x = U, M, D and y = L, C, R in the first period.

Now Period 1 Consider the strategies s12(h1) = M if h1 = (U,L) and D otherwise & s22(h1) = C if h1 = (U,L) and R otherwise. With these strategies the players’ payoffs for the game starting in period 1 are: which yields a subgame perfect equilibrium with cooperative, Non-Nash stage game play in period 1!

What about infinite repetition? First, two definitions: Feasible Payoff: Any convex combination of the pure strategy profile payoffs. i is feasible if i = sA s ui(s) where s  0 for s  A and sA s = 1. Average Payoff : (1 - ) t = 1  t - 1 ui(ht) where 1    0 is the discount factor. Theorem (Friedman 1971): Let G be a finite, static game of complete information. Let (e1,…,eN) denote the payoffs from a Nash equilibrium of G, and let (x1,…,xN) denote any other feasible payoffs from G. If xi > ei for every i and if  is sufficiently close to one, then there exists a subgame perfect Nash equilibrium of the infinitely repeated game G that achieves (x1,…,xN) as the average payoff. Often referred to as the Folk Theorem, but there are now lots of different versions of this Folk Theorem.

What does this result mean? In infinitely repeated games, we can get lots of subgame perfect equilibria. These equilibria can include actions in a stage game that are not Nash equilibrium actions for that stage game. You can get cooperative behavior in a Prisoner’s Dilemma! Lets see what I mean.

Consider the Prisoner’s Dilemma Consider the strategy: Play C in Period 1, Play C in period t > 1 if at’ = (C, C) for all t’  t, Otherwise play D. Can we find a discount rate such that this strategy is subgame perfect for this Prisoner’s Dilemma if it is repeated infinitely?

The answer to this question is yes! Suppose Player j is playing this type of strategy. At any point in time, Player j has either chosen D in the past in response to i’s choice of D or he has always chosen C because i has always chosen C. So, we must consider whether the strategy above is a best response for player i under both of these circumstances.

If D has been chosen in the past, player j will always choose D in the future. What is optimal for i now will be optimal for i in the future due to infinite repetition. Let VC & VD be the current value of playing strategy C & D. If C is optimal, i’s payoff from here on out will be VC = 0 +  VC such that VC = 0. If D is optimal, i’s payoff from here on out will be VD = 1 + VD such that VD = 1/(1 -  ). VD > VC, so D is optimal.

If D has not been chosen in the past, player j will choose C in the immediate future and will continue to do so as long as i does. But if i chooses D, j will follow suit from here on out. Again, what is optimal for i now will be optimal for i in the future due to infinite repetition. If C is optimal, i’s payoff from here on out will be VC = 2 +  VC such that VC = 2/(1 -  ). If D is optimal, i’s payoff from here on out will be VD = 3 + /(1 - ). VC >/=/< VD when  >/=/< ½.

To summarize As long as  > ½, this strategy will constitute a subgame perfect Nash strategy for the infinitely repeated Prisoner’s Dilemma. This type of strategy is often referred to as a trigger strategy. Bad behavior on the part of one player triggers bad behavior on the part of his opponent from here on after. Are there other trigger strategies that can work? YES!

General Trigger Strategy Define i*: equilibrium payoffs (per stage) iD: defection payoff iP: punishment payoffs (Nash equilibrium payoff per stage) Assume iD > i* > iP Cheating doesn’t pay when: or Are there other types of strategies that can work? YES! LOTS MORE!

So what are we to make of all this? It does provide an explanation for cooperation in games where cooperation seems unlikely. However, the explanation tells us that almost anything is possible. So, what type of behavior can we expect? The theory provides few answers. There has been a lot of research on repeated Prisoner Dilemma games to understand the best way to play as well as how people actually play. Of particular interest is Axelrod (1984). Axelrod had researches submit various strategies and had computers play them to see which ones performed the best. Tit-for-Tat strategies tended to perform the best the best.

Application: Cournot Duopoly with Repeated Play Who are the players? Two Firms Who can do what when? Firms Choose Output Each Period (qit for i = 1,2) to Infinity & Beyond Who knows what when? Firms Know Output Choices for all Previous Periods How are players rewarded based on what they do?

Stage Game Output & Profit Cournot Nash Equilibrium Output q1C = q2C = qC = a/3 Profit 1C = 2C = C = a2/9 Collusive Monopoly Outcome Output q1M = q2M = qM = a/4 Profit 1M = 2M = M = a2/8 Is it possible to sustain the collusive Monopoly outcome as a subgame perfect Nash equilibrium with infinite repetition?

Consider the Strategy Period 1: qi1 = qM Period t > 1: qit = qM if qit’ = qjt’ = qM for t’ < t qit = qC otherwise

Lets check to see if this proposed strategy is a subgame perfect Nash equilibrium. To accomplish this, we need to show that the strategy is a Nash equilibrium in all possible subgames. Our task is simplified here by the fact that there are only two distinct types of subgames: qit’ ≠ qM or qjt’ ≠ qM for some t’ < t qit’ = qjt’ = qM for all t’ < t

First consider qit’ ≠ qM or qjt’ ≠ qM for some t’ < t With this history, the proposed strategy says both players should choose qC. So, lets see what the optimal output in period t is for Firm i given Firm j will always choose qC.

Firm i’s optimal strategy is to choose the Cournot output just like the proposed strategy says!

Now consider qit’ = qjt’ = qM for some t’ < t With this history, the proposed strategy says both players should choose qM. So, lets see what the optimal output in period t is for Firm i given Firm j will always choose qM as long as Firm i chooses qM.

First, suppose that Firm i chooses qM in period t and forever after.

Now, suppose Firm i choose something other than qM in period t. Recall that we have already solved the optimization problem for which implies qis = qC for all s > t and

such that

a subgame perfect Nash equilibrium in this infinitely repeated game! Finally, Firm i will prefer to choose the Monopoly output forever after if or Therefore, if the discount rate is high enough, the proposed strategy will constitute a subgame perfect Nash equilibrium in this infinitely repeated game!

Is this the only subgame perfect Nash equilibrium? Hardly! One criticism of trigger strategies like our proposed strategy is that they do not permit cooperation to be reestablished. It is possible to find subgame perfect Nash equilibrium strategies that allow cooperation to be reestablished: Period 1: qi1 = qM. Period t > 1: qit = qM if qjt – 1 = qM or qjt – 1 = x x otherwise Though defining and proving such strategies are subgame perfect can be an arduous task!