Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.

Slides:



Advertisements
Similar presentations
Chapter 17: Making Complex Decisions April 1, 2004.
Advertisements

Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Game Theory Assignment For all of these games, P1 chooses between the columns, and P2 chooses between the rows.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Mixed Strategies CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte.
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
Joint Strategy Fictitious Play Sherwin Doroudi. “Adapted” from J. R. Marden, G. Arslan, J. S. Shamma, “Joint strategy fictitious play with inertia for.
ECO290E: Game Theory Lecture 5 Mixed Strategy Equilibrium.
Game Theory and Computer Networks: a useful combination? Christos Samaras, COMNET Group, DUTH.
Calibrated Learning and Correlated Equilibrium By: Dean Foster and Rakesh Vohra Presented by: Jason Sorensen.
An Introduction to... Evolutionary Game Theory
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Zero-Sum Games (follow-up)
Regret Minimization and the Price of Total Anarchy Paper by A. Blum, M. Hajiaghayi, K. Ligett, A.Roth Presented by Michael Wunder.
Learning in games Vincent Conitzer
Network Theory and Dynamic Systems Game Theory: Mixed Strategies
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
Introduction to Nash Equilibrium Presenter: Guanrao Chen Nov. 20, 2002.
1 Duke PhD Summer Camp August 2007 Outline  Motivation  Mutual Consistency: CH Model  Noisy Best-Response: QRE Model  Instant Convergence: EWA Learning.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
1 Economics & Evolution Number 2. 2 Reading List.
Convergent Learning in Unknown Graphical Games Dr Archie Chapman, Dr David Leslie, Dr Alex Rogers and Prof Nick Jennings School of Mathematics, University.
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
Static Games of Complete Information: Equilibrium Concepts
Beyond Nash: Raising the Flag of Rebellion Yisrael Aumann University of Haifa, 11 Kislev 5771 ( )
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Review Midterm3/19 3/12.
Advanced Microeconomics Instructors: Wojtek Dorabialski & Olga Kiuila Lectures: Mon. & Wed. 9:45 – 11:20 room 201 Office hours: Mon. & Wed. 9:15 – 9:45.
1 Computing Nash Equilibrium Presenter: Yishay Mansour.
Learning in Games. Fictitious Play Notation! For n Players we have: n Finite Player’s Strategies Spaces S 1, S 2, …, S n n Opponent’s Strategies Spaces.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
Rationality and information in games Jürgen Jost TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA A A A AAA Max Planck.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
QR 38, 2/22/07 Strategic form: dominant strategies I.Strategic form II.Finding Nash equilibria III.Strategic form games in IR.
Nash Equilibrium - definition A mixed-strategy profile σ * is a Nash equilibrium (NE) if for every player i we have u i (σ * i, σ * -i ) ≥ u i (s i, σ.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Review Midterm3/23 3/9.
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
1 On the Agenda(s) of Research on Multi-Agent Learning by Yoav Shoham and Rob Powers and Trond Grenager Learning against opponents with bounded memory.
1 Economics & Evolution. 2 Cournot Game 2 players Each chooses quantity q i ≥ 0 Player i’s payoff is: q i (1- q i –q j ) Inverse demand (price) No cost.
1 Issues on the border of economics and computation נושאים בגבול כלכלה וחישוב Congestion Games, Potential Games and Price of Anarchy Liad Blumrosen ©
Game Theory Statistics 802. Lecture Agenda Overview of games 2 player games representations 2 player zero-sum games Render/Stair/Hanna text CD QM for.
Minimax strategies, Nash equilibria, correlated equilibria Vincent Conitzer
CPS Learning in games Vincent Conitzer
Game Theory.
MAKING COMPLEX DEClSlONS
NOBEL WP Szept Stockholm Game Theory in Inter-domain Routing LÓJA Krisztina - SZIGETI János - CINKLER Tibor BME TMIT Budapest,
Derivative Action Learning in Games Review of: J. Shamma and G. Arslan, “Dynamic Fictitious Play, Dynamic Gradient Play, and Distributed Convergence to.
Learning in Multiagent systems
Introduction to Game Theory application to networks Joy Ghosh CSE 716, 25 th April, 2003.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
Mixed Strategies and Repeated Games
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Game Theory (Microeconomic Theory (IV)) Instructor: Yongqin Wang School of Economics, Fudan University December, 2004.
Vincent Conitzer CPS Learning in games Vincent Conitzer
5.1.Static Games of Incomplete Information
Replicator Dynamics. Nash makes sense (arguably) if… -Uber-rational -Calculating.
Game theory basics A Game describes situations of strategic interaction, where the payoff for one agent depends on its own actions as well as on the actions.
Replicator Dynamics.
Game Theory in Wireless and Communication Networks: Theory, Models, and Applications Lecture 2 Bayesian Games Zhu Han, Dusit Niyato, Walid Saad, Tamer.
Chapter 14 & 15 Repeated Games.
Chapter 14 & 15 Repeated Games.
Vincent Conitzer Learning in games Vincent Conitzer
Vincent Conitzer CPS Learning in games Vincent Conitzer
Presentation transcript:

Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley

Concave, Learning, Cooperative Concave Games Learning in Games Cooperative Games

Concave Games Motivation In many applications, the possible actions belong to a continuous set. For instance one chooses prices, transmission rates, or power levels. In such situations, one specifies reward functions instead of a matrix or rewards. We explain results on Nash equilibria for such games

Concave Games: Preliminaries Many situations are possible: No NE 3 NE1 NE J.B. Rosen, “ Existence and Uniqueness of Equilibrium Points for Concave N-Person Games,” Econometrica, 33, , July 1965

Concave Game Definition: Concave Game Definition: Nash Equilibrium

Concave Game Theorem: Existence Proof

Concave Game Specialized Case:

Concave Game Definition: Diagonally Strictly Concave

Concave Game Theorem: Uniqueness

Concave Game Theorem: Uniqueness - Bilinear Case:

Concave Game Local Improvements

Learning in Games Motivation Examples Models Fictitious Play Stochastic Fictitious Play Fudenberg D. and D.K. Levine (1998), The Theory of Learning in Games. MIT Press, Cambridge, Massachusetts. Chapters 1, 2, 4.

Motivation Explain equilibrium as result of players “learning” over time (instead of as the outcome of fully rational players with complete information)

Examples: 1 Fixed Player Model If P1 is patient and knows P2 chooses her play based on her forecast of P1’s plays, then P1 should always play U to lead P2 to play R A sophisticated and patient player who faces a naïve opponent can develop a reputation for playing a fixed strategy and obtain the rewards of a Stackelberg leader Large Population Models Most of the theory avoids possibility above by assuming random pairings in a large population of anonymous users In such a situation, P1 cannot really teach much to the rest of the population, so that myopic play (D, L) is optimal Naïve play: Ignore that you affect other players’ strategies LR U1, 03, 2 D2, 14, 0

Examples: 2 Cournot Adjustment Model Each player selects best response to other player’s strategy in previous period Converges to unique NE in this case This adjustment is a bit naïve … 1 2 BR2 BR1 NE

Models Learning Model: Specifies rules of individual players and examines their interactions in repeated game Usually: Same game is repeated (some work on learning from similar games) Fictitious Play: Players observe result of their own match, play best response to the historical frequency of play Partial Best-Response Dynamics: In each period, a fixed fraction of the population switches to a best response to the aggregate statistics from the previous period Replicator Dynamics: Share of population using each strategy grows at a rate proportional to that strategy’s current payoff

Fictitious Play Each player computes the frequency of the actions of the other players (with initial weights) Each player selects best response to the empirical distribution (need not be product) Theorem: Strict NE are absorbing for FP If s is a pure strategy and is steady-state for FP, then s = NE Proof: Assume s(t) = s = strict NE. Then, with a := a(t) …, p(t+1) = (1 – a)p(t) + a  (s), so that u(t+1, r) = (1 – a)u(p(t), r) + au(  (s), r), which is maximized by r = s if u(p(t), r) is maximized by r = s. Converse: If converges, this means players do not want to deviate, so limit must be NE…

Fictitious Play Assume initial weights (1.5, 2) and (2, 1.5). Then (T, T)  (1.5, 3), (2, 2.5)  (T, H), (T, H)  (H, H), (H, H), (H, H)  (H, T)… Theorem: If under FP empirical converge, then product converges to NE Proof: If strategies converge, this means players do not want to deviate, so limit must be NE… Theorem: Under FP, empirical converge if one of the following holds 2x2 with generic payoffs Zero-sum Solvable by iterated strict dominance … Note: Empirical distributions need not converge HT H1, -1-1, 1 T 1, -1

Fictitious Play Assume initial weights (1, ) for P1 and P2. Then (A, A)  (2, )  (B, B)  (A, A)  (B, B)  (A, A), etc Empirical frequencies converge to NE However, players get 0 Correlated strategies, not independent (Fix: Randomize …) AB A0, 01, 1 B 0, 0

Stochastic Fictitious Play Motivation: Avoid discontinuity in FP Hope for a stronger form of convergence: not only of the marginals, but also of the intended plays

Stochastic Fictitious Play Definitions: Reward of i = u(i, s) + n(i, s i ), n has positive support on interval BR(i,  )(s i ) = P[n(i, s i ) is s.t. s i = BR to  ] Nash Distribution: if  i = BR(i,  ), all i Harsanyi’s Purification Theorem: For generic payoffs, ND  NE if support of perturbation  0. Key feature: BR is continuous and close to original BR. Definitions: Reward of i = u(i, s) + n(i, s i ), n has positive support on interval BR(i,  )(s i ) = P[n(i, s i ) is s.t. s i = BR to  ] Nash Distribution: if  i = BR(i,  ), all i Harsanyi’s Purification Theorem: For generic payoffs, ND  NE if support of perturbation  0. Key feature: BR is continuous and close to original BR. Matching Pennies

Stochastic Fictitious Play Theorem (Fudenberg and Kreps, 93): Assume 2x2 game has unique mixed NE If smoothing is small enough, then NE is globally stable for SFP Theorem (K&Y 95, B&H 96) Assume 2x2 game has unique strict NE The unique intersection of smoothed BR is a global attractor for SFP Assume 2x2 game has 2 strict NE and one mixed NE. The SFP converges to one of the strict NE, w.p. 1. Note: Cycling is possible for SFP in multi-player games Theorem (Fudenberg and Kreps, 93): Assume 2x2 game has unique mixed NE If smoothing is small enough, then NE is globally stable for SFP Theorem (K&Y 95, B&H 96) Assume 2x2 game has unique strict NE The unique intersection of smoothed BR is a global attractor for SFP Assume 2x2 game has 2 strict NE and one mixed NE. The SFP converges to one of the strict NE, w.p. 1. Note: Cycling is possible for SFP in multi-player games

Stochastic Fictitious Play Other justification for randomization: Protection against opponent’s mistakes Learning rules should be Safe: average utility ≥ minmax Universally consistent: utility ≥ utility if frequency were known but not order of plays Randomization can achieve universal consistency (e.g., SFP)

Stochastic Fictitious Play Stimulus-Response (Reinforcement learning): Increase probability of plays that give good results General observation: It is difficult to discriminate learning models on the basis of experimental data: SFP, SR, etc. seem all about comparable

Cooperative Games Motivation Notions of Equilibrium Nash Bargaining Equilibrium Shapley Value

Cooperative Games: Motivation The Nash equilibriums may not be the most desirable outcome for the players. Typically, players benefit by cooperating. We explore some notions of equilibrium that players achieve under cooperation.

Cooperative Games: Equilibria

Cooperative Games: Nash B.E. Definition: Nash Bargaining Equilibrium Fact Interpretation:

Cooperative Games: Nash B.E. Example: NBE: Social:

Cooperative Games: Nash B.E. Axiomatic Justification  At NE, sum of relative increases is zero.

Shapley Value Example: Shapley Value:

Fixed Point Theorems Theorem (Brower):

Brower Labels (1, 3)Labels (2, 3) One path through doors (1, 2) must end up in triangle (1, 2, 3). [Indeed: Odd number of boundary doors.]

Brower Take small triangle (1, 2, 3) Divide it into triangles as before; it contains another (1, 2, 3); Continue in this way. Pick z(n) in triangle n. Let z = lim z(n). Claim: f(z) = z. Proof: If f(z) is not z, then z(n) and f(z(n)) are in different small triangles at stage n; but then z(n) cannot be in a (1, 2, 3) triangle ….

Notes on Kakutani Theorem:

Kakutani