Multiagent Interactions

Slides:



Advertisements
Similar presentations
Concepts of Game Theory I
Advertisements

Concepts of Game Theory II. 2 The prisioners reasoning… Put yourself in the place of prisoner i (or j)… Reason as follows: –Suppose I cooperate… If j.
Chapter 17: Making Complex Decisions April 1, 2004.
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Game Theory Assignment For all of these games, P1 chooses between the columns, and P2 chooses between the rows.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Game Theory S-1.
SIF8072 Distributed Artificial Intelligence and Intelligent Agents
Game Theory “Доверяй, Но Проверяй” - Russian Proverb (Trust, but Verify) - Ronald Reagan Mike Shor Lecture 6.
Game Theory “Доверяй, Но Проверяй” (“Trust, but Verify”) - Russian Proverb (Ronald Reagan) Topic 5 Repeated Games.
Infinitely Repeated Games. In an infinitely repeated game, the application of subgame perfection is different - after any possible history, the continuation.
Non-Cooperative Game Theory To define a game, you need to know three things: –The set of players –The strategy sets of the players (i.e., the actions they.
6-1 LECTURE 6: MULTIAGENT INTERACTIONS An Introduction to MultiAgent Systems
1 Chapter 14 – Game Theory 14.1 Nash Equilibrium 14.2 Repeated Prisoners’ Dilemma 14.3 Sequential-Move Games and Strategic Moves.
The basics of Game Theory Understanding strategic behaviour.
1 Game Theory. By the end of this section, you should be able to…. ► In a simultaneous game played only once, find and define:  the Nash equilibrium.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
Coye Cheshire & Andrew Fiore March 21, 2012 // Computer-Mediated Communication Collective Action and CMC: Game Theory Approaches and Applications.
Short introduction to game theory 1. 2  Decision Theory = Probability theory + Utility Theory (deals with chance) (deals with outcomes)  Fundamental.
Part 3: The Minimax Theorem
Game-theoretic analysis tools Necessary for building nonmanipulable automated negotiation systems.
Games What is ‘Game Theory’? There are several tools and techniques used by applied modelers to generate testable hypotheses Modeling techniques widely.
Prisoner’s Dilemma. The scenario In the Prisoner’s Dilemma, you and Lucifer are picked up by the police and interrogated in separate cells without the.
An Introduction to Game Theory Part I: Strategic Games
2008/02/06Lecture 21 ECO290E: Game Theory Lecture 2 Static Games and Nash Equilibrium.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
5/16/20151 Game Theory Game theory was developed by John Von Neumann and Oscar Morgenstern in Economists! One of the fundamental principles of.
Eponine Lupo.  Game Theory is a mathematical theory that deals with models of conflict and cooperation.  It is a precise and logical description of.
Slide 1 of 13 So... What’s Game Theory? Game theory refers to a branch of applied math that deals with the strategic interactions between various ‘agents’,
Chapter 12 Choices Involving Strategy McGraw-Hill/Irwin Copyright © 2008 by The McGraw-Hill Companies, Inc. All Rights Reserved.
Algoritmi per Sistemi Distribuiti Strategici
Game-Theoretic Approaches to Multi-Agent Systems Bernhard Nebel.
A Memetic Framework for Describing and Simulating Spatial Prisoner’s Dilemma with Coalition Formation Sneak Review by Udara Weerakoon.
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
Review: Game theory Dominant strategy Nash equilibrium
Design of Multi-Agent Systems Teacher Bart Verheij Student assistants Albert Hankel Elske van der Vaart Web site
Introduction to Game Theory and Behavior Networked Life CIS 112 Spring 2009 Prof. Michael Kearns.
QR 38, 2/22/07 Strategic form: dominant strategies I.Strategic form II.Finding Nash equilibria III.Strategic form games in IR.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
Minimax strategies, Nash equilibria, correlated equilibria Vincent Conitzer
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
A Game-Theoretic Approach to Strategic Behavior. Chapter Outline ©2015 McGraw-Hill Education. All Rights Reserved. 2 The Prisoner’s Dilemma: An Introduction.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Introduction Giovanni Neglia.
Chapters 29, 30 Game Theory A good time to talk about game theory since we have actually seen some types of equilibria last time. Game theory is concerned.
The Science of Networks 6.1 Today’s topics Game Theory Normal-form games Dominating strategies Nash equilibria Acknowledgements Vincent Conitzer, Michael.
Lecture 5 Introduction to Game theory. What is game theory? Game theory studies situations where players have strategic interactions; the payoff that.
Section 2 – Ec1818 Jeremy Barofsky
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Strategic Behavior in Business and Econ Static Games of complete information: Dominant Strategies and Nash Equilibrium in pure and mixed strategies.
Mohsen Afsharchi Multiagent Interaction. What are Multiagent Systems?
1 Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A. Franz J. Kurfess CPE/CSC 580: Intelligent Agents 1.
Lec 23 Chapter 28 Game Theory.
The Prisoner’s Dilemma or Life With My Brother and Sister John CT.
Taylor. Na, Amy. Hong, Brian. Sang, Luck Yoon. What is the Game Theory? intelligent rational decision-makers Game theory is "the study of mathematical.
Advanced Subjects in GT Outline of the tutorials Static Games of Complete Information Introduction to games Normal-form (strategic-form) representation.
Game theory basics A Game describes situations of strategic interaction, where the payoff for one agent depends on its own actions as well as on the actions.
Introduction to Game Theory
Oligopoly & Game Theory Lecture 27
Decision Theory and Game Theory
Multiagent Systems Game Theory © Manfred Huber 2018.
LECTURE 6: MULTIAGENT INTERACTIONS
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Normal Form (Matrix) Games
Presentation transcript:

Multiagent Interactions A game theoretical approach Matteo Venanzi bvenanzi@gmail.com

Outline Utilities and preferences Multiagent encounters Solution concepts The Prisoner’s Dilemma IPD in a social network

The spheres of influence

Utilities and preferences Self-interested agents: Each agent has its own preferences and desires over the states of the world (non-cooperative game theory). Modelling preferences: Outcomes (states of the world): Ω = {w1, w2, ...} Utility function: ui: Ω → R

Preference ordering Utility functions lead to preference orderings over outcomes Preference over w Strict preference over w Properties Reflexivity Transitivity Comparability

Multiagent Encounter Interaction as a game: the states of the world can be seen as the outcomes of a game. Assume we have just two agents (players) Ag={i,j}, The final outcome in Ω depends on the combination of actions selected by each agent. State transformer function:

Multiagent Encounter Normal-form game (or strategic-form game): A strategic interaction is familiarly represented in game theory as a tuple (N, A, u), where: N is (finite) the set of player. A = A1 x A2 x ... x An where Ai is the set of actions available to player i U = (u1, u2, ... , un) utility functions of each player Payoff matrix: n-dimentional matrix with cells enclosing the values of the utility functions for each player i cooperates i defects j cooperates 1, 1 1, 4 j defects 4,1 4, 4

Given a game scenario, how a rational agent will act? Multiagent Encounter Rational agent’s strategy: Given a game scenario, how a rational agent will act? Apparently: “just” maximize the expected payoff (single-agent point of view). In most cases unfeasible because the individual best strategy depends on the choices of others (multi-agent point of view) In practice: answer in solution concepts: Dominant strategies Pareto optimality Nash equilibrium

Solution Concepts Best response: given the player j’s strategy sj, the player i’s best response to sj is the strategy si that gives the highest payoff for player i. i: C i: D j: C 1, 1 1, 4 j: D 4,4 4, 1 Example: j plays C -----> i’s best response = D j plays D -----> i’s best response = C

Dominated strategy can be removed from the table Solution Concepts Dominant strategy: A strategy si* is dominant for player i if no matter what strategy sj agent j chooses, i will do at least as well playing si* as it would doing anything else. (si* is dominant if it is the best response to all of agent j’s strategies.) i: C i: D j: C 1, 4 1, 1 j: D 4,1 4, 4 Example: - D is the dominant strategy for player j There are not dominant strategies for player i Dominated strategy can be removed from the table

Dominated strategy can be removed from the table Solution Concepts Dominant strategy: A strategy si* is dominant for player i if no matter what strategy sj agent j chooses, i will do at least as well playing si* as it would doing anything else. (si* is dominant if it is the best response to all of agent j’s strategies.) i: C i: D j: C 1, 4 1, 1 j: D 4,1 4, 4 i: C i: D j: C 1, 4 1, 1 j: D 4,1 4, 4 Example: - D is the dominant strategy for player j There are not dominant strategies for player i Dominated strategy can be removed from the table

Dominated strategy can be removed from the table Solution Concepts Dominant strategy: A strategy si* is dominant for player i if no matter what strategy sj agent j chooses, i will do at least as well playing si* as it would doing anything else. (si* is dominant if it is the best response to all of agent j’s strategies.) i: C i: D j: D 4,1 4, 4 Example: - D is the dominant strategy for player j There are not dominant strategies for player i Dominated strategy can be removed from the table

Solution Concepts Pareto optimality (or Pareto efficiency): An outcome for the game is said to be Pareto optimal (or Pareto efficient) if there is no other outcome that makes one agent better off without making another agent worse off. If an outcome w is not Pareto optimal, then there is another outcome w’ that makes everyone as happy, if not happier, than w. “Reasonable” agents would agree to move to w’ in this case. (Even if I don’t directly benefit from w’, you can benefit without me suffering.)

Solution Concepts Pareto optimality (or Pareto efficiency): An outcome for the game is said to be Pareto optimal (or Pareto efficient) if there is no other outcome that makes one agent better off without making another agent worse off. Girl: whole cake Girl: half cake Girl: nothing Boy: whole cake Boy: Half cake Boy: Nothing If an outcome w is not Pareto optimal, then there is another outcome w’ that makes everyone as happy, if not happier, than w. “Reasonable” agents would agree to move to w’ in this case. (Even if I don’t directly benefit from w’, you can benefit without me suffering.)

Solution Concepts Pareto optimality (or Pareto efficiency): An outcome for the game is said to be Pareto optimal (or Pareto efficient) if there is no other outcome that makes one agent better off without making another agent worse off. Girl: whole cake Girl: half cake Girl: nothing Boy: whole cake Boy: Half cake Boy: Nothing Not Pareto efficient Not Pareto efficient Pareto efficient Not Pareto efficient Pareto efficient Not Pareto efficient Pareto efficient Not Pareto efficient Not Pareto efficient If an outcome w is not Pareto optimal, then there is another outcome w’ that makes everyone as happy, if not happier, than w. “Reasonable” agents would agree to move to w’ in this case. (Even if I don’t directly benefit from w’, you can benefit without me suffering.) Pareto optimality is a nice proberty but not very useful for selecting strategies

Nash Equilibrium Nash equilibrium for pure strategies: Two strategies s1 and s2 are in Nash equilibrium if: under the assumption that agent i plays s1, agent j can do no better than play s2; AND under the assumption that agent j plays s2, agent i can do no better than play s1. Neither agent has any incentive to deviate from a Nash equilibrium. Nash equilibrium represents the “Rational” outcome of a game played by self-interested agents. Unfortunately: Not every interaction scenario has a Nash equilibrium. Some interaction scenarios have more than one Nash equilibrium. But: Nash’s Theorem (1951): Every game with a finite number of players and a finite set of possible strategies has at least one Nash equilibrium in mixed strategies. However: “The complexity of finding a Nash equilibrium is the most important concrete open question on the boundary of P today” [Papadimitriou, 2001] John Forbes Nash, Jr. Bluefield. 1928 Nobel Prize for Economics, 1994 “A beautiful mind”

Finding Nash Equilibria Battle of the sexes game: husband and wife wish to go to the movies and they are undecided between “lethal Weapon (LW)” and “Wondrous Love (WL)”. They much prefer to go together rather than separate to the movie, altought the wife prefers LW and the husband prefers (WL)”. husband: LW husband: WL wife: LW 2, 1 0, 0 wife: WL 0,0 1, 2 Easy way to find pure-strategy Nash equilibria in a payoff matrix: Take the cell where the first number is the maximum of the column and check if the second number is the maximum of that row.

Finding Nash Equilibria Battle of the sexes game: husband and wife wish to go to the movies and they are undecided between “lethal Weapon (LW)” and “Wondrous Love (WL)”. They much prefer to go together rather than separate to the movie, altought the wife prefers LW and the husband prefers (WL)”. husband: LW husband: WL wife: LW 2, 1 0, 0 wife: WL 0,0 1, 2 Easy way to find pure-strategy Nash equilibria in a payoff matrix: Take the cell where the first number is the maximum of the column and check if the second number is the maximum of that row.

Finding Nash Equilibria Max-min strategy: the maxmin strategy for player i is the strategy si that maximizes the i’s payoff in the worst case. (Minmax is the dual strategy) Example: (zero-sum game: the sum of the utilities is always zero) Player 2: A Player 2: B Player 1: A 2, -2 3, -3 Player 1: B 0,0 4, -4 From the player 1’s point of view: - If he plays A, the min payoff is 2. - If he plays B, the min payoff is 0. Player 1 chooses A. With the same reasoning: player 2 chooses A too. The “maxmin solution” is (A, A). It is also the (only) Nash equilibrium of the game. In fact: Minmax Theorem ( von Neuman, 1928 ) : In any finite, two-players, zero-sum game maxmin strategies, minmax strategies and Nash equilibria coincide.

The Prisoner’s Dilemma Two men are collectively charged with a crime and held in separate cells, with no way of meeting or communicating. They are told that: if one confesses and the other does not, the confessor will be freed, and the other will be jailed for five years; if both confess, then each will be jailed for two years. Both prisoners know that if neither confesses, then they will each be jailed for one year. Well-known (and ubiquitous) puzzle in game theory. The fundamental paradox of multi-agent interactions. Payoff matrix for prisoner’s dilemma: Player 2 confesses Player 2 does not confess Player 1 confesses 1, 1 5, 0 Player 1 does not confess 0,5 3, 3

The Prisoner’s Dilemma Player 2 Defects Player 2 Cooperates Player 1 Defects 1, 1 5, 0 Player 1 Cooperates 0,5 3, 3 (Confession = defection, No confession = cooperation) Solution concepts: D is a dominant strategy. (D, D) is the only Nash equilibrium. All outcomes except (D;D) are Pareto optimal. Game theorist’s conclusions: In the worst case Defection guarantees a highest payoff so Defection is for both players the rational move to make. Both agents defects and get payoff = 2 Why dilemma..? They both know they could make better off by both cooperating and each get payoff =3. Cooperation is too dangerous!! In the one-shot prisoner’s dilemma the individual rational choice is defection

Iterated Prisoner’s Dilemma Observation: If the PD is repeated a number of time (so, I will meet my opponent again) then the incentive to Defect appears to decrease and Cooperation can finally emerge (Iterated Prisoner’s Dilemma). Axelrod in 1985 organized two computer tournament, inviting participants to submit their strategies to play the IPD. The length of the game was unknown to the players. 15 strategies submitted. Best strategies: SPITEFUL: Cooperate until receive a defection, then always defect MISTRUST: Defect, then play the last opponent’s move. TIT FOR TAT: Cooperate on 1st round, then play what the opponent played on the previous round. TIT_FOR_TAT won both the tournaments!!! (Still today, nobody was able to “clearly” defeat TFT)

Axelrod’s Results Even though TFT was the best strategy in the Axelrod’s tournament, it doesn’t mean it the Optimal Strategy for the IPD. In fact: Axelrod’s Theorem: If the game is long enough and the players do care about their future encounters, a universal optimal strategy for the Iterated Prisoner’s Dilemma does not exist. Axelrod also suggests the following rules for a successful strategy for the IPD: Don’t be envious: Remember the PD is not a zero-sum game. Don’t play as you would strike your opponent! Be nice: Never be the first to defect. Retaliate appropriately: There must be a proper measure in rewarding cooperation and punishing defection. Don’t be too clever: The performance of a strategy are not necessarily related to its complexity.

IPD in a Social Network Imagine agents playing the IPD, arranged as a graph where the edges represents rating relationships. The agent can use the knowledge and experience of his neighbours for retrieving ratings about his opponent and inferring his reputation. Rating propagation: Mechanism to propagate ratings from one node to another within the network. possible solution: the propagated rating the product of the weights on the edges along the path connecting the two players. In case of Multiple Path, take the one with the highest weight product. Strategy: Cooperate with probability Pmax

IPD in a Social Network Random Network TrustRate-Agents get all around the same score The ranking keeps the features of the Axelrod’s Tournament

IPD in a Social Network Dense Graph Increases the number of ratings. There is always a nice path between any couple of players.

Agent 11 inserted at round 55/100 IPD in a Social Network Get new agents into the game in progress Agent 11 inserted at round 55/100 The network helps the new player to immediately recognize the opponents with who cooperate or to avoid. Agent 11 scores the same points than the starting players.

IPD in a Social Network Conclusions: Acquaintance’s network allows to promote cooperation in the IPD with a distributed approach. No data overload on a single agent. Information spread through the whole network. Incoming agents are immediately in conditions to get into the logics of cooperation. If surrounded by some nice neighbourhood, the agent is able to immediately distinguish cooperative players from defective players, even without any prior experience.

Thanks! Questions..?

References An introduction to Multiagent Systems (2nd edition). M. Wooldridge, John Wiley, 2009. [Cap. 11] Multiagent Systems. Algorithmic, Game-Theoretic, and Logical Foundations. Y. Shoham, K. Leyton-Brown, Cambridge, 2009. [Cap. 3] (Free download at http://www.masfoundations.org) Trust, Kinship and Locality in the Iterated Prisoner’s Dilemma. M. Venanzi. MSc Thesis. [Cap. 5] (PDF available here)