THE PRICE OF STOCHASTIC ANARCHY Christine ChungUniversity of Pittsburgh Katrina LigettCarnegie Mellon University Kirk PruhsUniversity of Pittsburgh Aaron.

Slides:



Advertisements
Similar presentations
You have been given a mission and a code. Use the code to complete the mission and you will save the world from obliteration…
Advertisements

Analysis of Computer Algorithms
Copyright © 2008 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics, 9e Managerial Economics Thomas Maurice.
Generating Random Spanning Trees Sourav Chatterji Sumit Gulwani EECS Department University of California, Berkeley.
Subspace Embeddings for the L1 norm with Applications Christian Sohler David Woodruff TU Dortmund IBM Almaden.
STATISTICS Sampling and Sampling Distributions
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Energy-Efficient Distributed Algorithms for Ad hoc Wireless Networks Gopal Pandurangan Department of Computer Science Purdue University.
Thursday, March 7 Duality 2 – The dual problem, in general – illustrating duality with 2-person 0-sum game theory Handouts: Lecture Notes.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Title Subtitle.
My Alphabet Book abcdefghijklm nopqrstuvwxyz.
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
SUBTRACTING INTEGERS 1. CHANGE THE SUBTRACTION SIGN TO ADDITION
Addition Facts
Year 6 mental test 5 second questions
Vote Elicitation with Probabilistic Preference Models: Empirical Estimation and Cost Tradeoffs Tyler Lu and Craig Boutilier University of Toronto.
Comp 122, Spring 2004 Order Statistics. order - 2 Lin / Devi Comp 122 Order Statistic i th order statistic: i th smallest element of a set of n elements.
Discrete time Markov Chain
Randomized Algorithms Randomized Algorithms CS648 1.
Recurrences : 1 Chapter 3. Growth of function Chapter 4. Recurrences.
ABC Technology Project
Hash Tables.
Online Algorithm Huaping Wang Apr.21
Cache and Virtual Memory Replacement Algorithms
5-1 Chapter 5 Theory & Problems of Probability & Statistics Murray R. Spiegel Sampling Theory.
1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Joint work with Andreas Krause 1 Daniel Golovin.
Weighted moving average charts for detecting small shifts in process mean or trends The wonders of JMP 1.
© Jude Shavlik 2006, David Page 2007 CS 760 – Machine Learning (UW-Madison)RL Lecture, Slide 1 Reinforcement Learning (RL) Consider an “agent” embedded.
Squares and Square Root WALK. Solve each problem REVIEW:
Variants of Nash Equilibrium CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte 1.
© 2012 National Heart Foundation of Australia. Slide 2.
Lets play bingo!!. Calculate: MEAN Calculate: MEDIAN
Reaching Agreements II. 2 What utility does a deal give an agent? Given encounter  T 1,T 2  in task domain  T,{1,2},c  We define the utility of a.
Chapter 5 Test Review Sections 5-1 through 5-4.
GG Consulting, LLC I-SUITE. Source: TEA SHARS Frequently asked questions 2.
Addition 1’s to 20.
25 seconds left…...
Week 1.
We will resume in: 25 Minutes.
Local Search Jim Little UBC CS 322 – CSP October 3, 2014 Textbook §4.8
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) Oct, 5, 2012.
A SMALL TRUTH TO MAKE LIFE 100%
PSSA Preparation.
Oct 9, 2014 Lirong Xia Hypothesis testing and statistical decision theory.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 14 From Randomness to Probability.
James McGalliard, FEDSIM CMG Southern Region Raleigh - April 11, 2014 Richmond – April 17,
Peter Key, Laurent Massoulie, Don Towsley Infocom 07 presented by Park HoSung 1 Path selection and multipath congestion control.
Auction Theory Class 5 – single-parameter implementation and risk aversion 1.
Probabilistic Reasoning over Time
NON - zero sum games.
Price Of Anarchy: Routing
This Segment: Computational game theory Lecture 1: Game representations, solution concepts and complexity Tuomas Sandholm Computer Science Department Carnegie.
Evolution and Repeated Games D. Fudenberg (Harvard) E. Maskin (IAS, Princeton)
Game-theoretic analysis tools Necessary for building nonmanipulable automated negotiation systems.
The Evolution of Conventions H. Peyton Young Presented by Na Li and Cory Pender.
A camper awakens to the growl of a hungry bear and sees his friend putting on a pair of running shoes, “You can’t outrun a bear,” scoffs the camper. His.
Nov 2003Group Meeting #2 Distributed Optimization of Power Allocation in Interference Channel Raul Etkin, Abhay Parekh, and David Tse Spectrum Sharing.
Load Balancing, Multicast routing, Price of Anarchy and Strong Equilibrium Computational game theory Spring 2008 Michal Feldman.
MAKING COMPLEX DEClSlONS
Learning in Multiagent systems
1 Efficiency and Nash Equilibria in a Scrip System for P2P Networks Eric J. Friedman Joseph Y. Halpern Ian Kash.
Game-theoretic analysis tools Tuomas Sandholm Professor Computer Science Department Carnegie Mellon University.
Price of Anarchy Georgios Piliouras. Games (i.e. Multi-Body Interactions) Interacting entities Pursuing their own goals Lack of centralized control Prediction?
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
Equilibrium transitions in stochastic evolutionary games Dresden, ECCS’07 Jacek Miękisz Institute of Applied Mathematics University of Warsaw.
Replicator Dynamics. Nash makes sense (arguably) if… -Uber-rational -Calculating.
Presented By Aaron Roth
Presentation transcript:

THE PRICE OF STOCHASTIC ANARCHY Christine ChungUniversity of Pittsburgh Katrina LigettCarnegie Mellon University Kirk PruhsUniversity of Pittsburgh Aaron RothCarnegie Mellon University

Time Needed Machine 1 Machine 2 Job 1 Job 2 Job 3 Load Balancing on Unrelated Machines 2 Machine 1 Machine 2 n players, each with a job to run, chooses one of m machines to run it on Each players goal is to minimize her jobs finish time. NOTE: finish time of a job is equal to load on the machine where the job is run.

Time Needed Machine 1 Machine 2 Job 1 Job 2 Job 3 Load Balancing on Unrelated Machines 3 Machine 1 Machine 2 n players, each with a job to run, chooses one of m machines to run it on Each players goal is to minimize her jobs finish time. NOTE: finish time of a job is equal to load on the machine where the job is run.

n players, each with a job to run, chooses one of m machines to run it on Each players goal is to minimize her jobs finish time. NOTE: finish time of a job is equal to load on the machine where the job is run. Time Needed Machine 1 Machine 2 Job 1 Job 2 Job 3 Load Balancing on Unrelated Machines 4 Machine 1 Machine 2

Time Needed Machine 1 Machine 2 Job 1 Job 2 Job 3 Load Balancing on Unrelated Machines 5 Machine 1 Machine 2 n players, each with a job to run, chooses one of m machines to run it on Each players goal is to minimize her jobs finish time. NOTE: finish time of a job is equal to load on the machine where the job is run.

Unbounded Price of Anarchy in the Load Balancing Game on Unrelated Machines 6 Price of Anarchy (POA) measures the cost of having no central authority. Let an optimal assignment under centralized authority be one in which makespan is minimized. POA = (makespan at worst Nash)/(makespan at OPT) Bad POA instance: 2 players and 2 machines (L and R). OPT here costs δ. Worst Nash costs 1. Price of Anarchy: L job 2 job 1 R δ 1 1 δ δ 1

Drawbacks of Price of Anarchy A solution characterization with no road map. If there is more than one Nash, dont know which one will be reached. Strong assumptions must be made about the players: e.g., fully informed and fully convinced of one anothers rationality. Nash are sometimes very brittle, making POA results feel overly pessimistic. 7

Evolutionary Game Theory 8 Young (1993) specified a model of adaptive play.

Evolutionary Game Theory 9 Young (1993) specified a model of adaptive play that allows us to predict which solutions will be chosen in the long run by self-interested decision-making agents with limited info and resources. I dispense with the notion that people fully understand the structure of the games they play, that they have a coherent model of others behavior, that they can make rational calculations of infinite complexity, and that all of this is common knowledge. Instead I postulate a world in which people base their decisions on limited data, use simple predictive models, and sometimes do unexplained or even foolish things. – P. Young, Individual Strategy and Social Structure, 1998

Evolutionary Game Theory 10 Young (1993) specified a model of adaptive play. Adaptive play allows us to predict which solutions will be chosen in the long run by self-interested decision-making agents with limited info and resources.

In each round of play, each player uses some simple, reasonable dynamics to decide which strategy to play. E.g., imitation dynamics Sample s of the last mem strategies I played Play the strategy whose average payoff was highest (breaking ties uniformly at random) best response dynamics Sample the other players realized strategy in s of the last mem rounds. Assume this sample represents the probability distribution of what the other player will play the next round, and play a strategy that is a best response (minimizes my expected cost). Adaptive Play Example 11 L job 2 job 1 R δ 1 1 δ

In each round of play, each player uses some simple, reasonable dynamics to decide which strategy to play. E.g., imitation dynamics Sample s of the last mem strategies I played Play the strategy whose average payoff was highest (breaking ties uniformly at random) best response dynamics Sample the other players realized strategy in s of the last mem rounds. Assume this sample represents the probability distribution of what the other player will play the next round, and play a strategy that is a best response (minimizes my expected cost). Adaptive Play Example 12 L job 2 job 1 R δ 1 1 δ

Let mem = 4. If s = 3, each player randomly samples three past plays from the memory, and picks the strategy among them that worked best (yielded the highest payoff). LLRR LLLL Adaptive Play Example: a Markov process 13 LLLL player 1 player 2 3/4 1/4 1 LLLR LLLL LLLR RRRR LRRR RRRR... LLRL LLLL (Then there are 2^8 = 256 total states in the state space.) 1 L job 2 job 1 R δ 1 1 δ

Absorbing Sets of the Markov Process 14 An absorbing set is a set of states that are all reachable from one another, but cannot reach any states outside of the set. In our example, we have 4 absorbing sets: But which state we end up in depends on our initial state. Hence we perturb our Markov process as follows: During each round, each player, with probability ε, does not use imitation dynamics, but instead chooses a machine at random. 1 RRRR 1 LLLL 1 RRRR 1 LLLL NASH OPT L job 2 job 1 R δ 1 1 δ

Stochastic Stability 15 The perturbed process has only one big absorbing set (any state is reachable from any other state). Hence we have a unique stationary distribution μ ε (where μ ε P = μ ε ). The probability distribution μ ε is the time-average asymptotic frequency distribution of P ε. A state z is stochastically stable if L job 2 job 1 R δ 1 1 δ

Finding Stochastically Stable States 16 Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. 1 RRRR 1 LLLL 1 RRRR 1 LLLL L job 2 job 1 R δ 1 1 δ

Finding Stochastically Stable States 17 Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. RRRR LLLL RRRR LLLL 1 RRRR LRLR LRLR LRLR LRLR LLLL RRRRL LLL 3 L job 2 job 1 R δ 1 1 δ

Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. = cost of min spanning tree rooted there Finding Stochastically Stable States 18 RRRR LLLL RRRR LLLL L job 2 job 1 R δ 1 1 δ

Finding Stochastically Stable States 19 Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. RRRR LLLL RRRR LLLL L job 2 job 1 R δ 1 1 δ 6 = cost of min spanning tree rooted there

Finding Stochastically Stable States 20 Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. RRRR LLLL RRRR LLLL L job 2 job 1 R δ 1 1 δ 6 = cost of min spanning tree rooted there

Finding Stochastically Stable States 21 Theorem (Young, 1993): The stochastically stable states are those states contained in the absorbing sets of the unperturbed process that have minimum stochastic potential. RRRR LLLL RRRR LLLL Stochastically Stable! L job 2 job 1 R δ 1 1 δ 6 = cost of min spanning tree rooted there

Recap: Adaptive Play Model Assume the game is played repeatedly by players with limited information and resources. Use a decision rule (aka learning behavior or selection dynamics) to model how each player picks her strategy for each round. This yields a Markov Process where the states represent fixed-sized histories of game play. Add noise (players make mistakes with some small positive probability and dont always behave according to the prescribed dynamics) 22

23 Stochastic Stability The states in the perturbed Markov process with positive probability in the long-run are the stochastically stable states (SSS). In our paper, we define the Price of Stochastic Anarchy (PSA) to be 23

Recall bad instance: POA = 1/ δ (unbounded) But the bad Nash in this case is not a SSS. In fact, OPT is the only SSS here. So PSA = 1 in this instance. Our main result: For the game of load balancing on unrelated machines, while POA is unbounded, PSA is bounded. Specifically, we show PSA m(Fib (n) (mn+1)), which is m times the (mn+1)th n-step Fibonacci number. We also exhibit instances of the game where PSA > m. PSA for Load Balancing (m is the number of machines, n is the number of jobs/players) Ω(m) PSA mFib (n) (mn+1) 24 L job 2 job 1 R δ 1 1 δ

In the game of load balancing on unrelated machines, we found that while POA is unbounded, PSA is bounded. Indeed, in the bad POA instances for many games, the worst Nash are not stochastically stable. Finding PSA in these games are interesting open questions that may yield very illuminating results. PSA allows us to determine relative stability of equilibria, distinguishing those that are brittle from those that are more robust, giving us a more informative measure of the cost of having no central authority. Closing Thoughts 25

You might notice in this game that if players could coordinate or form a team, they would play OPT. Instead of being unbounded, [AFM2007] have shown the strong price of anarchy is O(m). We conjecture that PSA is also O(m), i.e., that a linear price of anarchy can be achieved without player coordination. Conjecture 26 L job 2 job 1 R δ 1 1 δ