Why How We Learn Matters Russell Golman Scott E Page.

Slides:



Advertisements
Similar presentations
Holling 1973.
Advertisements

Innuendos and Common Knowledge. Consider the following examples of innuendos: "I hear you're the foreman of the jury in the Soprano trial. It's an important.
Ultimatum Game Two players bargain (anonymously) to divide a fixed amount between them. P1 (proposer) offers a division of the “pie” P2 (responder) decides.
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Game Theory Assignment For all of these games, P1 chooses between the columns, and P2 chooses between the rows.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Mixed Strategies CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte.
3. Basic Topics in Game Theory. Strategic Behavior in Business and Econ Outline 3.1 What is a Game ? The elements of a Game The Rules of the.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Congestion Games with Player- Specific Payoff Functions Igal Milchtaich, Department of Mathematics, The Hebrew University of Jerusalem, 1993 Presentation.
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
The basics of Game Theory Understanding strategic behaviour.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
Nash Equilibria By Kallen Schwark 6/11/13 Fancy graphs make everything look more official!
Common Welfare, Strong Currencies and the Globalization Process Esteban Guevara Hidalgo Center for Nonlinear and Complex Systems (Como, Italy) ICTP-TRIL.
Learning in games Vincent Conitzer
Game-theoretic analysis tools Necessary for building nonmanipulable automated negotiation systems.
Maynard Smith Revisited: Spatial Mobility and Limited Resources Shaping Population Dynamics and Evolutionary Stable Strategies Pedro Ribeiro de Andrade.
1 Evolution & Economics No. 5 Forest fire p = 0.53 p = 0.58.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Algoritmi per Sistemi Distribuiti Strategici
An Introduction to Game Theory Part II: Mixed and Correlated Strategies Bernhard Nebel.
A Memetic Framework for Describing and Simulating Spatial Prisoner’s Dilemma with Coalition Formation Sneak Review by Udara Weerakoon.
Communication Networks A Second Course Jean Walrand Department of EECS University of California at Berkeley.
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
APEC 8205: Applied Game Theory Fall 2007
Nash Q-Learning for General-Sum Stochastic Games Hu & Wellman March 6 th, 2006 CS286r Presented by Ilan Lobel.
Distributed Computing with Adaptive Heuristics Michael Schapira Princeton Innovations in Computer Science 09 January 2011 Partially supported by NSF Aaron.
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
Learning in Games Scott E Page Russell Golman & Jenna Bednar University of Michigan-Ann Arbor.
Quantal Response Equilibrium APEC 8205: Applied Game Theory Fall 2007.
QR 38, 2/22/07 Strategic form: dominant strategies I.Strategic form II.Finding Nash equilibria III.Strategic form games in IR.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
Quantitative Reasoning 38 The Strategy of International Politics Professor Lisa Martin Head TF Yev Kirpichevsky TF Dilyan Donchev.
Games of Incomplete Information. These games drop the assumption that players know each other’s preferences. Complete info: players know each other’s preferences.
CPS Learning in games Vincent Conitzer
Common Knowledge: The Math. We need a way to talk about “private information”
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
Learning in Multiagent systems
Nash equilibrium Nash equilibrium is defined in terms of strategies, not payoffs Every player is best responding simultaneously (everyone optimizes) This.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
1 Economics & Evolution Number 3. 2 The replicator dynamics (in general)
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Dynamic Games & The Extensive Form
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
Topic 3 Games in Extensive Form 1. A. Perfect Information Games in Extensive Form. 1 RaiseFold Raise (0,0) (-1,1) Raise (1,-1) (-1,1)(2,-2) 2.
Ásbjörn H Kristbjörnsson1 The complexity of Finding Nash Equilibria Ásbjörn H Kristbjörnsson Algorithms, Logic and Complexity.
Mixed Strategies and Repeated Games
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Strategic Behavior in Business and Econ Static Games of complete information: Dominant Strategies and Nash Equilibrium in pure and mixed strategies.
Replicator Dynamics. Nash makes sense (arguably) if… -Uber-rational -Calculating.
By: Donté Howell Game Theory in Sports. What is Game Theory? It is a tool used to analyze strategic behavior and trying to maximize his/her payoff of.
Evolution of Cooperation in Mobile Ad Hoc Networks Jeff Hudack (working with some Italian guy)
Chapter 12 Game Theory Presented by Nahakpam PhD Student 1Game Theory.
Game theory Chapter 28 and 29
A useful reduction (SAT -> game)
Game theory basics A Game describes situations of strategic interaction, where the payoff for one agent depends on its own actions as well as on the actions.
Q 2.1 Nash Equilibrium Ben
The Duality Theorem Primal P: Maximize
Project BEST Game Theory.
A useful reduction (SAT -> game)
Game theory Chapter 28 and 29
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Presentation transcript:

Why How We Learn Matters Russell Golman Scott E Page

Overview BIG Picture Game Theory Basics Nash Equilibrium –Equilibrium something about the other. –Stability –Basins Learning Rules Why Learning Matters

Big Question If you had one word to describe the social, political, and economic worlds around you would you choose equilibrium or complex?

Methodological Question Do we construct simple, illustrative, insight generating models (PD, Sandpile, El Farol) or do we construct high fidelity, realistic models?

My Answer: BOTH! Both types of models are useful in their own right. In addition, each tells us something about the other.

My Answer: BOTH! Knowledge of simple models helps us construct better high fidelity models. Large models show if insights from simple models still apply.

Today’s Exercise How does how agents learn influence outcomes?

Step Way Back Complex Adaptive System - Agents - Variation - Selection*

Examples Best respond to current state Better respond Mimic best Mimic better Include portions of best or better Random with death of the unfit

Equilibrium Science We can start by looking at the role that learning rules play in equilibrium systems. This will give us some insight into whether they’ll matter in complex systems.

Game Theory Players Actions Payoffs

Players

Actions Cooperate: C Defect: D

Payoffs 4,40,6 6,02,2 C D C D

Best Responses 4,40,6 6,02,2 C D C D

Best Responses 4,40,6 6,02,2 C D C D

Nash Equilibrium 4,40,6 6,02,2 C D C D 2,2

“Equilibrium” Based Science Step 1: Set up game Step 2: Solve for equilibrium Step 3: Show how equilibrium depends on parameters of model Step 4: Provide empirical support

Is Equilibrium Enough? Existence: Equilibrium exists Stability: Equilibrium is stable Attainable: Equilibrium is attained by a learning rule.

Stability Stability can only be defined relative to a learning dynamic. In dynamical systems, we often take that dynamic to be a best response function, but with human actors we need not assume people best respond.

Existence Theorem Theorem: Finite number of players, finite set of actions, then there exists a Nash Equilibrium. Pf: Show best response functions are upper hemi continuous and then apply Kakutani’s fixed point theorem

Battle of Sexes Game 3,10,0 1,3 EF CG EF CG

Three Equilibria 3,10,0 1,3 1/4 3/4 3/4 1/4

Unstable Mixed? 3,10,0 1,3 1/4 +e 3/4 - e EF CG 3/4 + 3e 3/4 - 3e

Note the Implicit Assumption Our stability analysis assumed that Player 1 would best respond to Player 2’s tremble. However, the learning rule could be go to the mixed strategy equilibrium. If so, Player 1 would sit tight and Player 2 would return to the mixed strategy equilibrium.

Empirical Foundations We need to have some understanding of how people learn and adapt to say anything about stability.

Classes of Learning Rules Belief Based Learning Rules: People best respond given their beliefs about how other people play. Replicator Learning Rules: People replicate successful actions of others.

Examples Belief Based Learning Rules: Best response functions Replicator Learning Rules: Replicator dynamics

Stability Results An extensive literature provides conditions (fairly week) under which the two learning rules have identical stability property. Synopsis: Learning rules do not matter

Basins Question Do games exist in which best response dynamics and replicator dynamics produce very different basins of attraction? Question: Does learning matter?

Best Response Dynamics x = mixed strategy of Player 1 y = mixed strategy of Player 2 dx/dt = BR(y) - x dy/dt = BR(x) - y

Replicator Dynamics dx i /dt = x i (  i -  ave ) dy i /dt = y i (  i -  ave )

Symmetric Matrix Game A B C ABC

Conceptualization Imagine a large population of agents playing this game. Each chooses an action. The population distribution over actions creates a mixed strategy. We can then study the dynamics of that population given different learning rules.

Best Response Basins A B C ABC A > B iff 60p A + 60p B + 30p C > 30p A + 70p B + 20p C A > C iff 60p A + 60p B + 30p C > 70p A + 25p B + 35p C

Best Response Basins A A B C C B CB

Stable Equilibria A A B C C B CB

Best Response Basins A A B C C B AB

Replicator Dynamics A A B C C B ?B

Replicator Dynamics Basins A A B C C B ABB

Recall: Basins Question Do games exist in which best response dynamics and replicator dynamics produce very different basins of attraction? Question: Does learning matter?

Conjecture For any  > 0, There exists a symmetric matrix game such that the basins of attraction for distinct equilibria under continuous time best response dynamics and replicator dynamics overlap by less than 

Results Thm 1 (SP): Can be done if number of actions goes to infinity

Results Thm 1 (SP): Can be done if number of actions goes to infinity Thm 2 (RG): Can be done if number of actions scales with 1/ 

Results Thm 1 (SP): Can be done if number of actions goes to infinity Thm 2 (RG): Can be done if number of actions scales with 1/  Thm 3 (RG): Cannot be done with two actions.

Results Thm 1 (SP): Can be done if number of actions goes to infinity Thm 2 (RG): Can be done if number of actions scales with 1/  Thm 3 (RG): Cannot be done with two actions. Thm 4 (SP): Can be done with four!

Collective Action Game SI Coop Pred Naive N N2N2 00-N 2 0 SI Coop Pred Naive

Intuition: Naïve Goes Away SI Coop Pred

Intuition: Naïve Goes Away SI Coop Pred Coop SI

Best Response SI Coop Pred Coop SI

Best Response Pred Coop SI

Replicator SI Coop Pred Coop SI

Collective Action Game SI Coop Pred Naive N N2N2 00-N 2 0 SI Coop Pred Naive

The Math dx i /dt = x i (  i -  ave )  ave = 2x S + x C (1+Nx C ) dx c /dt = x c [(1+ Nx C- - 2x S - x C (1+Nx C )] dx c /dt = x c [(1+ Nx C )(1- x C ) - 2x S ]

Choose N > 1/  Assume x c >  dx c /dt = x c [(1+ Nx C )(1- x C ) - 2x S ] dx c /dt >  [2(1-  ) - 2(1-  )] = 0 Therefore, x c always increases.

Aside: Why Care? Replicator dynamics often thought of as being cultural learning. Best response learning thought of as self interested learning. Societies differ by degree of individualism. These results show that how society is structure could affect the ability to solve collective action problems.

Results (Cont’d) Conjecture (SP): There does not exist a game with three actions such that the basins have vanishing overlap.

Results (Cont’d) Conjecture (SP): There does not exist a game with three actions such that the basins have vanishing overlap. Thm 5 (RG): There does exist a game with three actions such that the basins have vanishing overlap

Genericity of Results Example Proof for a functional form Proof for a class of functions General Result

What do we have? Examples: 3 and 4 dimensions General Result: 3 dimensions is minimal.

Another General Result Recall that in the (very cool) four dimensional example, that initially predatory behavior was a best response with probability one. Moreover, it was not an equilibrium. Turns out, this is always true!!

Theorem: In any symmetric matrix game for which best response and replicator dynamics attain different equilibria with probability one, there exists an action A that is both an initial best response with probability one and is not an equilibrium.

From Science to Art Insight: If I’m constructing a large scale ABM and some actions will win for a short time and then die off, then I had better experiment with lots of learning rules.

Two Aggregation Questions Q1: What if I take an average of the learning rules? Q2: What if some people use one rule and some use another? Do I get the same result as everyone used the same hybrid rule?

Two Aggregation Questions Q1: What if I take an average of the learning rules? Anything you want Q2: What if some people use one rule and some use another? Do I get the same result as everyone used the same hybrid rule? No

Why Complexity?

Complex Issues Global Inequality Crime/Education Health Care Ecosystem Management Global Climate Change International Relations/Terrorism Epidemics

Needs More “thinking tools” PD, Sandpile, etc.. More science and improved art for constructing high fidelity models.