Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Slides:



Advertisements
Similar presentations
Introduction to Game Theory
Advertisements

9.1 Strictly Determined Games Game theory is a relatively new branch of mathematics designed to help people who are in conflict situations determine the.
GAME THEORY.
Module 4 Game Theory To accompany Quantitative Analysis for Management, Tenth Edition, by Render, Stair, and Hanna Power Point slides created by Jeff Heyl.
Name: Trương Hoài Anh Facebook: Quasar Hoaianh
APPENDIX An Alternative View of the Payoff Matrix n Assume total maximum profits of all oligopolists is constant at 200 units. n Alternative policies.
A very little Game Theory Math 20 Linear Algebra and Multivariable Calculus October 13, 2004.
Simultaneous- Move Games with Mixed Strategies Zero-sum Games.
Two-Player Zero-Sum Games
Operations Research Assistant Professor Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD4207 University of Palestine.
1 Chapter 4: Minimax Equilibrium in Zero Sum Game SCIT1003 Chapter 4: Minimax Equilibrium in Zero Sum Game Prof. Tsang.
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
Study Group Randomized Algorithms 21 st June 03. Topics Covered Game Tree Evaluation –its expected run time is better than the worst- case complexity.
Game theory.
Game Theory, Part 1 Game theory applies to more than just games. Corporations use it to influence business decisions, and militaries use it to guide their.
Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc
Part 3: The Minimax Theorem
Operations Research: Applications and Algorithms
Chapter 14 Game Theory to accompany Operations Research: Applications and Algorithms 4th edition by Wayne L. Winston Copyright (c) 2004 Brooks/Cole, a.
1 1 © 2003 Thomson  /South-Western Slide Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Matrices, Digraphs, Markov Chains & Their Use. Introduction to Matrices  A matrix is a rectangular array of numbers  Matrices are used to solve systems.
Operations Research: Applications and Algorithms
1. Markov Process 2. States 3. Transition Matrix 4. Stochastic Matrix 5. Distribution Matrix 6. Distribution Matrix for n 7. Interpretation of the Entries.
What is the probability that the great-grandchild of middle class parents will be middle class? Markov chains can be used to answer these types of problems.
Lectures in Microeconomics-Charles W. Upton Minimax Strategies.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Games of Chance Introduction to Artificial Intelligence COS302 Michael L. Littman Fall 2001.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 68 Chapter 9 The Theory of Games.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 60 Chapter 8 Markov Processes.
To accompany Quantitative Analysis for Management, 8e by Render/Stair/Hanna S-1 © 2003 by Prentice Hall, Inc. Upper Saddle River, NJ Supplement 1.
Game Theory.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Minimax Strategies. Everyone who has studied a game like poker knows the importance of mixing strategies. –With a bad hand, you often fold –But you must.
Game Theory Statistics 802. Lecture Agenda Overview of games 2 player games representations 2 player zero-sum games Render/Stair/Hanna text CD QM for.
Game Theory.
8.2 Regular Stochastic Matrices
MAKING COMPLEX DEClSlONS
Presentation by: H. Sarper
To accompany Quantitative Analysis for Management,9e by Render/Stair/Hanna M4-1 © 2006 by Prentice Hall, Inc. Upper Saddle River, NJ Module 4 Game.
9  Markov Chains  Regular Markov Chains  Absorbing Markov Chains  Game Theory and Strictly Determined Games  Games with Mixed Strategies Markov Chains.
Game Theory Part 2: Zero Sum Games. Zero Sum Games The following matrix defines a zero-sum game. Notice the sum of the payoffs to each player, at every.
The Design & Analysis of the Algorithms Lecture by me M. Sakalli Download two pdf files..
Chapter 14 Game Theory to accompany Operations Research: Applications and Algorithms 4th edition by Wayne L. Winston Copyright (c) 2004 Brooks/Cole, a.
1 1 Slide © 2006 Thomson South-Western. All Rights Reserved. Slides prepared by JOHN LOUCKS St. Edward’s University.
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Lecture 12. Game theory So far we discussed: roulette and blackjack Roulette: – Outcomes completely independent and random – Very little strategy (even.
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
Zero-sum Games The Essentials of a Game Extensive Game Matrix Game Dominant Strategies Prudent Strategies Solving the Zero-sum Game The Minimax Theorem.
Statistics Overview of games 2 player games representations 2 player zero-sum games Render/Stair/Hanna text CD QM for Windows software Modeling.
INDEX Introduction of game theory Introduction of game theory Significance of game theory Significance of game theory Essential features of game theory.
10.1 Properties of Markov Chains In this section, we will study a concept that utilizes a mathematical model that combines probability and matrices to.
Markov Games TCM Conference 2016 Chris Gann
GAME THEORY Day 5. Minimax and Maximin Step 1. Write down the minimum entry in each row. Which one is the largest? Maximin Step 2. Write down the maximum.
Copyright © 2006 Brooks/Cole, a division of Thomson Learning, Inc. Linear Programming: An Algebraic Approach 4 The Simplex Method with Standard Maximization.
9.2 Mixed Strategy Games In this section, we look at non-strictly determined games. For these type of games the payoff matrix has no saddle points.
Goldstein/Schnieder/Lay: Finite Math & Its Applications, 9e 1 of 60 Chapter 8 Markov Processes.
From DeGroot & Schervish. Example Occupied Telephone Lines Suppose that a certain business office has five telephone lines and that any number of these.
Game Theory [geym theer-ee] : a mathematical theory that deals with the general features of competitive situations in a formal abstract way.
Tools for Decision Analysis: Analysis of Risky Decisions
Input – Output Models P = # of items produced
Chapter 6 Game Theory (Module 4) 1.
Game Theory.
Game Theory II Solutions 1
Game Theory Day 4.
9.3 Linear programming and 2 x 2 games : A geometric approach
Operations Research: Applications and Algorithms
Systems of Linear Equations:
Operations Research: Applications and Algorithms
Presentation transcript:

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc 9 Markov Chains and the Theory of Games Markov Chains Regular Markov Chains Absorbing Markov Chains Game Theory and Strictly Determined Games Games with Mixed Strategies Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Any square matrix that satisfies the properties: 2. The sum of the entries in each column of T is 1. can be referred to as a stochastic matrix. Ex. Stochastic, columns add to 1 Not stochastic, column 1: sum = 1.3 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Markov Process (Chain) Stochastic Process in which the outcomes at any stage of the experiment depend only on the outcomes of the preceding stage. The state is the outcome at any stage. The outcome at the current stage is called the current state. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. We are given a process with 2 choices: A and B. It is expected that if a person chooses A then that person has a 30% probability of choosing A the next time. If a person chooses B then that person has a 60% probability of choosing B the next time. 0.3 A 0.4 A A B Basically: 0.7 B B 0.6 This can be represented by a transition matrix Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc State 1 State 2 The probability that something in state 1 will be in state 1 in the next step is a11 = 0.3 The probability that something in state 2 will be in state 1 in the next step is a12 = 0.4 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Transition Matrix A transition matrix associated with a Markov chain with n states is an n X n matrix T with entries aij Current State Next state Next state Current state 2. The sum of the entries in each column of T is 1. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. It has been found that of the people that eat brand X cereal, 85% will eat brand X again the next time and the rest will switch to brand Y. Also, 90% of the people that eat brand Y will eat brand Y the next time with the rest switching to brand X. At present, 70% of the people will eat brand X and the rest will eat brand Y. What percent will eat brand X after 1 cycle? X Initial state Transition: Y One cycle: So 62.5% will eat brand X Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Notice from the previous example: is called a distribution vector In general, the probability distribution of the system after n observations is given by Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Regular Markov Chain A stochastic matrix T is a regular Markov chain if the sequence T, T2, T3,… approaches a steady-state matrix in which the rows of the limiting matrix are all equal and all the entries are positive. Ex. Regular: all entries positive Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc T is Regular: all entries of T2 positive Ex. Notice Ex. Not regular: entries of T to a power will never all be positive. Notice Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Steady–State Distribution Vector Ex. Given Notice (after some work) Tends toward The steady-state distribution vector is the limiting vector from the repeated application of the transition matrix to the distribution vector. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Finding the Steady-State Distribution Vector Let T be a regular stochastic matrix. Then the steady-state distribution vector X may be found by solving the vector equation together with the condition that the sum of the elements of the vector X be equal to 1. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. Find the steady-state vector for the transition matrix: both are Also need: Which gives: So Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Absorbing Stochastic Matrix An absorbing stochastic matrix has the properties: There is at least one absorbing state. It is possible to go from any non-absorbing state to an absorbing state in one or more stages. Ex. Absorbing Matrix State 3 is an absorbing state and an object may go from state 2 or 1 (non-absorbing states) to state 3 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Given an absorbing stochastic matrix it is possible to rewrite it in the form: Absorbing Nonabsorbing I: identity matrix O: zero matrix Ex. 1 4 2 3 1 2 3 4 1 2 3 4 3 2 4 1 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Finding the Steady-State Matrix for an Absorbing Stochastic Matrix Suppose an absorbing stochastic matrix A has been partitioned into submatrices Then the steady-state matrix of A is given by Where the order of the identity matrix is chosen to have the same order as R. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. Compute the steady-state matrix for the matrix from the previous example: Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc The steady-state matrix is: 1 4 2 3 Original columns 3 2 4 1 Original rows This means that an object starting in state 3 will have a probability of 0.366 of being in state 2 in the long term. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Game Theory A combination of matrix methods with the theory of probability to determine the optimal strategies to be used when opponents are competing to maximize gains (minimize losses). Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. Rafe (row player) and Carley (column player) are playing a game where each holds out a red or black chip simultaneously (neither knows the other’s choice). The betting is summarized below Carley holds black Carley holds red Rafe holds black Carley pays Rafe $5 Rafe pays Carley $10 Rafe holds red Carley pays Rafe $2 Carley pays Rafe $3 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc We can summarize the game as a payoff matrix for Rafe: C1 C2 R1 R2 The game is a zero-sum game since one person’s payoff is the same as the other person’s loss. Rafe basically picks a row and Carley picks a column. Since the matrix is a payoff for Rafe, he wants to maximize the entry while Carley wants to minimize the entry. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Rafe should look at the minima for each row, then pick the larger of the minima. This is called the Maximin strategy. C1 C2 minima R1 R2 –10 2 maxima 5 3 Carley should look for the maxima of the columns, then pick the smallest of the maxima. This is called the Minimax strategy. From this we see that Rafe should pick row 2 while Carley should pick column 2. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Maxmin Strategy (R’s move) For each row (payoff matrix), find the smallest entry in that row. Choose the row for which the entry found in step 1 is as large as possible. Minmax Strategy (C’s move) For each column of the payoff matrix, find the largest entry in that column. Choose the column for which the entry found in step 1 is as small as possible. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. Determine the maximin and minimax strategies for each player in a game that has the payoff matrix: Row minima –2 –4 Column maxima 4 3 –2 The row player should pick row 1. The column player should pick column 3. *Note: the column player is favored under these strategies (win 2) Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Optimal Strategy The optimal strategy in a game is the strategy that is most profitable to a particular player. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Strictly Determined Game A strictly determined game is characterized by the following properties: There is an entry in the payoff matrix that is simultaneously the smallest entry in its row and the largest entry in its column. This entry is called the saddle point for the game. The optimal strategy for the row (column) player is precisely the maxmin (minmax) strategy and is the row (column) containing the saddle point. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc From a previous example: This is a strictly determined game with saddle point. The optimal strategies are for row player to pick row 1 and the column player to pick column 3. The value of the game is –2. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Mixed Strategies Making different moves during a game. A row (column) player may choose different rows (columns) during the game. Ex. The game below has no saddle point. From a minimax/maximin strategy the row player should pick row 2 and the column player should pick column 3. Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc For a mixed strategy, let the row player pick row 2, 80% of the time and row 1, 20 % of the time. Let the column player pick column 1, 2, and 3, 10%, 20%, and 70% of the time respectively. Column player: Row player: To find the expected value of the game we compute: payoff Expected value = 1.1 Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Expected Value of a Game Let P and Q be the mixed strategies for the row player R and the column player C respectively. The expected value, E, of the game is given by: Payoff matrix Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Optimal Strategies for Nonstrictly Determined Games Payoff matrix Row player: Column player: Where Value of the game: Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc

Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc Ex. Given the payoff matrix, find the optimal strategies and then the value of the game. Row player should pick each row 50% of the time. Column player should pick column 1, 90% of the time. Value: Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc