Satisfaction Equilibrium Stéphane Ross. Canadian AI 20062 / 21 Problem In real life multiagent systems :  Agents generally do not know the preferences.

Slides:



Advertisements
Similar presentations
Chapter 17: Making Complex Decisions April 1, 2004.
Advertisements

Infinitely Repeated Games
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Game Theory Assignment For all of these games, P1 chooses between the columns, and P2 chooses between the rows.
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Chapter Twenty-Eight Game Theory. u Game theory models strategic behavior by agents who understand that their actions affect the actions of other agents.
Tacit Coordination Games, Strategic Uncertainty, and Coordination Failure John B. Van Huyck, Raymond C. Battalio, Richard O. Beil The American Economic.
6-1 LECTURE 6: MULTIAGENT INTERACTIONS An Introduction to MultiAgent Systems
1 Chapter 14 – Game Theory 14.1 Nash Equilibrium 14.2 Repeated Prisoners’ Dilemma 14.3 Sequential-Move Games and Strategic Moves.
The basics of Game Theory Understanding strategic behaviour.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Chapter 6 Game Theory © 2006 Thomson Learning/South-Western.
Eponine Lupo.  Questions from last time  3 player games  Games larger than 2x2—rock, paper, scissors  Review/explain Nash Equilibrium  Nash Equilibrium.
Game Theory Advertising Example 1. Game Theory What is the optimal strategy for Firm A if Firm B chooses to advertise? 2.
Infinitely Repeated Games Econ 171. Finitely Repeated Game Take any game play it, then play it again, for a specified number of times. The game that is.
Course: Applications of Information Theory to Computer Science CSG195, Fall 2008 CCIS Department, Northeastern University Dimitrios Kanoulas.
Learning in games Vincent Conitzer
Network Theory and Dynamic Systems Game Theory: Mixed Strategies
Game-theoretic analysis tools Necessary for building nonmanipulable automated negotiation systems.
An Introduction to Game Theory Part I: Strategic Games
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
Non-cooperative Game Theory Notes by Alberto Bressan.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Algoritmi per Sistemi Distribuiti Strategici
IN SEARCH OF VALUE EQUILIBRIA By Christopher Kleven & Dustin Richwine xkcd.com.
Game-Theoretic Approaches to Multi-Agent Systems Bernhard Nebel.
Basics on Game Theory For Industrial Economics (According to Shy’s Plan)
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
Chapter Twenty-Eight Game Theory. u Game theory models strategic behavior by agents who understand that their actions affect the actions of other agents.
Lectures in Microeconomics-Charles W. Upton Game Theory.
XYZ 6/18/2015 MIT Brain and Cognitive Sciences Convergence Analysis of Reinforcement Learning Agents Srinivas Turaga th March, 2004.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
Abstract Though previous explorations of equilibria in game theory have incorporated the concept of error-making, most do not consider the possibility.
APEC 8205: Applied Game Theory Fall 2007
Nash Q-Learning for General-Sum Stochastic Games Hu & Wellman March 6 th, 2006 CS286r Presented by Ilan Lobel.
Coye Cheshire & Andrew Fiore June 28, 2015 // Computer-Mediated Communication Game Theory, Games, and CMC.
On Bounded Rationality and Computational Complexity Christos Papadimitriou and Mihallis Yannakakis.
Evolutionary Games The solution concepts that we have discussed in some detail include strategically dominant solutions equilibrium solutions Pareto optimal.
1 On the Agenda(s) of Research on Multi-Agent Learning by Yoav Shoham and Rob Powers and Trond Grenager Learning against opponents with bounded memory.
Simple search methods for finding a Nash equilibrium Ryan Porter, Eugene Nudelman, and Yoav Shoham Games and Economic Behavior, Vol. 63, Issue 2. pp ,
Strategic Game Theory for Managers. Explain What is the Game Theory Explain the Basic Elements of a Game Explain the Importance of Game Theory Explain.
Social Choice Session 7 Carmen Pasca and John Hey.
Chapter 12 Choices Involving Strategy Copyright © 2014 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written.
Exponential Moving Average Q- Learning Algorithm By Mostafa D. Awheda Howard M. Schwartz Presented at the 2013 IEEE Symposium Series on Computational Intelligence.
林偉楷 Taiwan Evolutionary Intelligence Laboratory.
General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning Duke University Machine Learning Group Discussion Leader: Kai Ni June 17, 2005.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Standard and Extended Form Games A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor, SIUC.
Presenter: Chih-Yuan Chou GA-BASED ALGORITHMS FOR FINDING EQUILIBRIUM 1.
Strategic Decisions in Noncooperative Games Introduction to Game Theory.
Dynamic Games & The Extensive Form
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Public Policy Analysis MPA 404 Lecture 24. Previous Lecture Graphical Analysis of Tariff and Quota Game Theory; The prisoner's Dilemma.
Section 2 – Ec1818 Jeremy Barofsky
1 What is Game Theory About? r Analysis of situations where conflict of interests is present r Goal is to prescribe how conflicts can be resolved 2 2 r.
Strategic Behavior in Business and Econ Static Games of complete information: Dominant Strategies and Nash Equilibrium in pure and mixed strategies.
Strategic Game Theory for Managers. Explain What is the Game Theory Explain the Basic Elements of a Game Explain the Importance of Game Theory Explain.
IJCAI’07 Emergence of Norms through Social Learning Partha Mukherjee, Sandip Sen and Stéphane Airiau Mathematical and Computer Sciences Department University.
ECO290E: Game Theory Lecture 3 Why and How is Nash Equilibrium Reached?
Lec 23 Chapter 28 Game Theory.
On the Difficulty of Achieving Equilibrium in Interactive POMDPs Prashant Doshi Dept. of Computer Science University of Georgia Athens, GA Twenty.
Game theory basics A Game describes situations of strategic interaction, where the payoff for one agent depends on its own actions as well as on the actions.
Microeconomics Course E
Introduction to Game Theory
Convergence, Targeted Optimality, and Safety in Multiagent Learning
Unit 4 SOCIAL INTERACTIONS.
Multiagent Systems Repeated Games © Manfred Huber 2018.
Molly W. Dahl Georgetown University Econ 101 – Spring 2009
Collaboration in Repeated Games
Normal Form (Matrix) Games
Presentation transcript:

Satisfaction Equilibrium Stéphane Ross

Canadian AI / 21 Problem In real life multiagent systems :  Agents generally do not know the preferences (rewards) of their opponents  Agents may not observe the actions of their opponents In this context, most game theoretic solution concepts are hardly applicable We may try to define equilibrium concepts that :  do not require complete information  are achievable through learning, over repeated play

Canadian AI / 21 Plan Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions

Canadian AI / 21 Presentation Plan Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions

Canadian AI / 21 Game model : Number of agents : Joint action space : Set of possible outcomes, the outcome function., agent i’s reward function. Agent only knows, and. After each turn, every agent observes an outcome.

Canadian AI / 21 Game model Observations:  The agents do not know the game matrix  They are unable to compute best responses and Nash Equilibrium.  They can only reason on their history of actions and rewards. Aa,?b,? Bc,?d,? a,b,c,d

Canadian AI / 21 Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions Presentation Plan

Canadian AI / 21 Since the agents can only reason on their history of payoff, we may adopt a satisfaction-based reasoning:  If an agent is satisfied by its current reward, it should keep playing the same strategy  An unsatisfied agent may decide to change its strategy according to some exploration function An equilibrium will arise when all agents are satisfied. Satisfaction Equilibrium

Canadian AI / 21 Formally :  is the satisfaction function of agent : if (agent i is satisfied) if (agent i is not satisfied)  is the satisfaction threshold of agent A joint strategy is a satisfaction equilibrium if :  Satisfaction Equilibrium

Canadian AI / 21 Example Prisoner’s dilemma Possible satisfaction matrix : CD C-1, -1-10, 0 D0, -10-8,-8 Dominant strategy : D Nash Equilibrium : (D,D) Pareto-Optimal : (C,C), (D,C), (C,D) CD C1, 1 0, 1 D 1, 00, 0 CD C1, 1 0, 1 D 1, 0 1, 1

Canadian AI / 21 Satisfaction Equilibrium However, even if a satisfaction equilibrium exists, it may be unreachable : ABC A1,1 0,1 B 1,0 0,1 C 1,00,11,0

Canadian AI / 21 Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions Presentation Plan

Canadian AI / 21 Satisfaction Equilibrium Learning If the satisfaction thresholds are fixed, we only need to apply the satisfaction-based reasoning:  Choose a strategy randomly  If satisfied, keep playing the same strategy  Else choose a new strategy randomly We can also use other exploration functions which favour actions that have not been explored often  Ex:

Canadian AI / 21 Satisfaction Equilibrium Learning We use a simple update rule:  When the agent is satisfied, we increment its satisfaction threshold by some variable  If the agent is unsatisfied, we decrement its satisfaction threshold of  is multiplied by a factor each turn such that it converges to 0 We also use a limited history of our previous satisfaction states and thresholds for each action to bound the value of the satisfaction threshold

Canadian AI / 21 Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions Presentation Plan

Canadian AI / 21 Results Fixed satisfaction thresholds  In simple games, we were always able to reach a satisfaction equilibrium.  Using a biased exploration improves the speed of convergence of the algorithm. Learning the satisfaction thresholds  We are generally able to learn the optimal satisfaction equilibrium in simple games.  Using a biased exploration improves the convergence percentage of the algorithm.  The factor and history size affects the convergence of the algorithm and need to be adjusted to get optimal results.

Canadian AI / 21 Results – Prisoner’s dilemma

Canadian AI / 21 Game model Satisfaction Equilibrium Satisfaction Equilibrium Learning Results Conclusion Questions Presentation Plan

Canadian AI / 21 Conclusion It is possible to learn stable outcomes without observing anything but our own rewards Satisfaction equilibria can be defined on any Pareto-Optimal solution.  However, satisfaction equilibria are not always reachable The proposed learning algorithms achieves good performance in simple games  However, they require game-specific adjustments for optimal performance

Canadian AI / 21 Conclusion For more information, you can consult my publications at:  Thank You!

Canadian AI / 21 Questions ?