Robust Mechanism Design with Correlated Distributions

Slides:



Advertisements
Similar presentations
(Single-item) auctions Vincent Conitzer v() = $5 v() = $3.
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
Game Theory in Wireless and Communication Networks: Theory, Models, and Applications Lecture 6 Auction Theory Zhu Han, Dusit Niyato, Walid Saad, Tamer.
Optimal auction design Roger Myerson Mathematics of Operations research 1981.
1 Regret-based Incremental Partial Revelation Mechanism Design Nathanaël Hyafil, Craig Boutilier AAAI 2006 Department of Computer Science University of.
Preference Elicitation Partial-revelation VCG mechanism for Combinatorial Auctions and Eliciting Non-price Preferences in Combinatorial Auctions.
Seminar In Game Theory Algorithms, TAU, Agenda  Introduction  Computational Complexity  Incentive Compatible Mechanism  LP Relaxation & Walrasian.
Yang Cai Oct 15, Interim Allocation rule aka. “REDUCED FORM” : Variables: Interim Allocation rule aka. “REDUCED FORM” : New Decision Variables j.
Bundling Equilibrium in Combinatorial Auctions Written by: Presented by: Ron Holzman Rica Gonen Noa Kfir-Dahav Dov Monderer Moshe Tennenholtz.
Planning under Uncertainty
Visual Recognition Tutorial
Agent Technology for e-Commerce Chapter 10: Mechanism Design Maria Fasli
Computational Criticisms of the Revelation Principle Vincent Conitzer, Tuomas Sandholm AMEC V.
Complexity of Mechanism Design Vincent Conitzer and Tuomas Sandholm Carnegie Mellon University Computer Science Department.
Automated Mechanism Design: Complexity Results Stemming From the Single-Agent Setting Vincent Conitzer and Tuomas Sandholm Computer Science Department.
AP Statistics Section 10.2 A CI for Population Mean When is Unknown.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
CPS 173 Mechanism design Vincent Conitzer
Sequences of Take-It-or-Leave-it Offers: Near-Optimal Auctions Without Full Valuation Revelation Tuomas Sandholm and Andrew Gilpin Carnegie Mellon University.
By: Amir Ronen, Department of CS Stanford University Presented By: Oren Mizrahi Matan Protter Issues on border of economics & computation, 2002.
Automated Design of Multistage Mechanisms Tuomas Sandholm (Carnegie Mellon) Vincent Conitzer (Carnegie Mellon) Craig Boutilier (Toronto)
Yang Cai Oct 08, An overview of today’s class Basic LP Formulation for Multiple Bidders Succinct LP: Reduced Form of an Auction The Structure of.
Regret Minimizing Equilibria of Games with Strict Type Uncertainty Stony Brook Conference on Game Theory Nathanaël Hyafil and Craig Boutilier Department.
Yang Cai Oct 06, An overview of today’s class Unit-Demand Pricing (cont’d) Multi-bidder Multi-item Setting Basic LP formulation.
Automated Mechanism Design Tuomas Sandholm Presented by Dimitri Mostinski November 17, 2004.
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 22.
Algorithmic Mechanism Design Shuchi Chawla 11/7/2001.
Automated mechanism design Vincent Conitzer
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Expressive Negotiation over Donations to Charities Vincent Conitzer and Tuomas Sandholm.
Bayesian Algorithmic Mechanism Design Jason Hartline Northwestern University Brendan Lucier University of Toronto.
Comp/Math 553: Algorithmic Game Theory Lecture 11
Automated mechanism design
Bayesian games and mechanism design
Applications of Automated Mechanism Design
Comp/Math 553: Algorithmic Game Theory Lecture 08
CPS Mechanism design Michael Albert and Vincent Conitzer
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
Computational Mechanism Design
Mechanism design with correlated distributions
Comp/Math 553: Algorithmic Game Theory Lecture 09
Non-Parametric Models
Vincent Conitzer CPS Repeated games Vincent Conitzer
CHAPTER 10 Comparing Two Populations or Groups
Implementation in Bayes-Nash equilibrium
ECO 173 Chapter 10: Introduction to Estimation Lecture 5a
When Security Games Go Green
Chapter 6. Large Scale Optimization
CONCEPTS OF ESTIMATION
More about Posterior Distributions
Implementation in Bayes-Nash equilibrium
Linear Programming Duality, Reductions, and Bipartite Matching
Vincent Conitzer Mechanism design Vincent Conitzer
Vincent Conitzer CPS 173 Mechanism design Vincent Conitzer
Automated mechanism design
Preference elicitation/ iterative mechanisms
LECTURE 21: CLUSTERING Objectives: Mixture Densities Maximum Likelihood Estimates Application to Gaussian Mixture Models k-Means Clustering Fuzzy k-Means.
Lecture 7 Sampling and Sampling Distributions
Vincent Conitzer Repeated games Vincent Conitzer
CHAPTER 10 Comparing Two Populations or Groups
Vincent Conitzer Computer Science Department
CS 416 Artificial Intelligence
CPS Preference elicitation/ iterative mechanisms
Weichao Mao, Zhenzhe Zheng, Fan Wu
Implementation in Bayes-Nash equilibrium
Chapter 6. Large Scale Optimization
Information, Incentives, and Mechanism Design
Vincent Conitzer CPS Repeated games Vincent Conitzer
Vincent Conitzer CPS Mechanism design Vincent Conitzer
CPS Bayesian games and their use in auctions
Presentation transcript:

Robust Mechanism Design with Correlated Distributions Michael Albert and Vincent Conitzer malbert@cs.duke.edu and conitzer@cs.duke.edu

Prior-Dependent Mechanisms In many situations we’ve seen, optimal mechanisms are prior dependent Myerson auction for independent bidder valuations Revenue maximization with efficient allocation with correlated bidder valuations (Cremer and McLean 1985; Albert, Conitzer, and Lopomo 2016) Strong budget balanced mechanisms with correlated valuations (Kosenok and Severinov 2008) There are different degrees of prior dependence What if we don’t know the prior? Or can only estimate it using past reports in an auction?

Painting LP Example We make reproductions of two paintings maximize 3x + 2y subject to 4x + 2y ≤ 16 x + 2y ≤ 8 x + y ≤ 5 x ≥ 0 y ≥ 0 Painting 1 sells for $30, painting 2 sells for $20 Painting 1 requires 4 units of blue, 1 green, 1 red Painting 2 requires 2 blue, 2 green, 1 red We have 16 units blue, 8 green, 5 red

Mis-Estimation of the Objective 2.4x + 2.6y maximize 3x + 2y subject to 4x + 2y ≤ 16 x + 2y ≤ 8 x + y ≤ 5 x ≥ 0 y ≥ 0 8 6 Estimated solution: x=2, y=3 4 optimal solution: x=3, y=2 Objective Value with Optimal Solution: 13 Objective Value with Estimated Solution: 12 2 2 4 6 8

Mis-Estimation of the Constraints maximize 3x + 2y subject to 4x + 2y ≤ 16 x + 2y ≤ 8 x + y ≤ 5 x ≥ 0 y ≥ 0 8 x + y ≤ 4.95 6 4 optimal solution: x=3, y=2 2 2 4 6 8

The Mechanism Design Problem Objective for mechanism design with optimal revenue 𝑚𝑎𝑥 p 𝜃,𝜔 ,x 𝜃,𝜔 𝜃,𝜔 ∈Θ×Ω 𝝅 𝜃,𝜔 x 𝜃,𝜔 Ex-Post Individual Rationality, for all θ∈Θ and ω∈Ω: p 𝜃,𝜔 𝑣 𝜃 −x 𝜃,𝜔 ≥0 Ex-Interim Individual Rationality, for all θ∈Θ: 𝜔 𝝅(𝜔|𝜃)(p 𝜃,𝜔 𝑣 𝜃 −x 𝜃,𝜔 ) ≥0 Ex-Post Incentive Compability, for all θ∈Θ and ω∈Ω: p 𝜃,𝜔 𝑣 𝜃 −x 𝜃,𝜔 ≥p 𝜃′,𝜔 𝑣 𝜃 −x 𝜃′,𝜔 BNE Incentive Compatibility, for all θ∈Θ: 𝜔 𝝅(𝜔|𝜃)(p 𝜃,𝜔 𝑣 𝜃 −x 𝜃,𝜔 ) ≥ 𝜔 𝝅(𝜔|𝜃)(p 𝜃′,𝜔 𝑣 𝜃 −x 𝜃′,𝜔 )

Consistent Set of Distributions Let 𝑃(𝐴) be the set of probability distributions over the set 𝐴. Then the space of all probability distributions over Θ×Ω can be represented as 𝑃(Θ×Ω). A subset 𝒫( 𝝅 )⊆𝑃(Θ×Ω) is a consistent set of distributions for the estimated distribution 𝝅 if the true distribution, 𝝅, is guaranteed to be in 𝒫( 𝝅 ) and 𝝅 ∈𝒫( 𝝅 ).

Something Between Ex-Post and Bayesian We need something between Ex-Post/Dominant Strategy and Interim/BNE. We need a prior dependent notion of incentive compatibility and individual rationality that is robust to small mis-estimations of the distributions. There is a catch: we can now allow the bidder to report the true distribution at the same time as he reports his valuation (Revelation Principle) A mechanism is a probability of allocation of the item, p 𝜃,𝛑,𝜔 , and a payment, x 𝜃,𝛑,𝜔 , that depends on the reported type, 𝜃, the reported distribution, 𝛑, and the external signal 𝜔. Define the bidder’s expected utility from reporting 𝜃′,𝛑′ when his true valuation is 𝜃,𝛑 and the external signal is 𝜔 as: 𝑈 𝜃,𝛑, 𝜃 ′ , 𝛑 ′ ,𝜔 =p 𝜃′,𝛑′,𝜔 𝑣 𝜃 −x 𝜃′,𝛑′,𝜔

Robust Individual Rationality and Incentive Compatibility A mechanism is robust individually rational for estimated bidder distribution 𝝅 and consistent set of distributions 𝒫( 𝝅 ) if for all 𝜃∈Θ and 𝛑∈𝒫 𝝅 : 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃,𝛑,𝜔) ≥0 A mechanism is robust incentive compatible for estimated bidder distribution 𝝅 and consistent set of distributions 𝒫( 𝝅 ) if for all 𝜃,𝜃′∈ Θ and 𝛑,𝛑′∈𝒫 𝝅 : 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃,𝛑,𝜔) ≥ 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃′,𝛑′,𝜔)

Should we listen to the bidder’s reported belief? The revelation principle still holds, and we can always construct a mechanism where the bidder reports his belief 𝛑 𝜔 𝜃 along with his type truthfully. We will only consider, with loss of generality, mechanisms that do not take the reported valuation, 𝛑, into account. A mechanism is a probability of allocation of the item, p 𝜃,𝜔 , and a payment, x 𝜃,𝜔 , that depends on the reported type, 𝜃, and the external signal, 𝜔. Define the bidder’s expected utility from reporting 𝜃′,𝛑′ when his true valuation and distribution is 𝜃,𝛑 and the external signal is 𝜔 as: 𝑈 𝜃,𝛑,𝜃′,𝜔 =p 𝜃′,𝜔 𝑣 𝜃 −x 𝜃′,𝜔 Note that this is not without loss of generality, but it reduces the set of payments and probabilities that must be specified to a finite set. Once we have a finite set, we can use techniques from automated mechanism design to design the optimal restricted robust mechanism.

Linear Program for Optimal Restricted Robust Mechanism 𝑚𝑎𝑥 p 𝜃,𝜔 ,x 𝜃,𝜔 𝜃,𝜔 ∈Θ×Ω 𝝅 𝜃,𝜔 x 𝜃,𝜔 Subject to: 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃,𝜔) ≥ 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈 𝜃,𝛑, 𝜃 ′ ,𝜔 ∀ 𝜃, 𝜃 ′ ∈Θ and ∀ 𝛑∈𝒫 𝝅 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃,𝜔) ≥0 ∀ 𝜃∈Θ and ∀ 𝛑∈𝒫 𝝅 0≤p 𝜃,𝜔 ≤1 ∀ 𝜃∈Θ and ∀ 𝜔∈Ω We have an infinite number of constraints!

Constraint Generation Linear Program (IR)For all 𝜃∈Θ: 𝑚𝑖𝑛 𝛑 𝜔∈Ω 𝛑 𝜔 𝜃 𝑈(𝜃,𝛑,𝜃,𝜔) (IC)For all 𝜃,𝜃′∈Θ: 𝑚𝑖𝑛 𝛑 𝜔∈Ω 𝛑 𝜔 𝜃 (𝑈(𝜃,𝛑,𝜃,𝜔) −𝑈(𝜃,𝛑,𝜃′,𝜔)) Subject to: 𝛑∈𝒫 𝝅 Solve the first linear program with a consistent set 𝒫 ′ 𝝅 ={ 𝝅 }. Solve each of the linear programs, and then add the solution, 𝛑, to 𝒫 ′ 𝝅 if the objective value is less than 0. Re-solve the original linear program and repeat. Guaranteed to terminate in a polynomial number of iterations (Kozlov, Tarasov, and Khachiyan 1980) What is the issue with this set of linear programs? We need the constraints to be finite!

Polyhedral Consistent Set If the set 𝒫 𝝅 can be characterized as an n-polyhedron, where n is polynomial in the number of bidder and external signal types, then the optimal robust mechanism can be computed in polynomial time. Example: Suppose that we restrict our attention to consistent sets such that for all 𝜃∈Θ and 𝜔∈Ω, 𝛑 𝜔 𝜃 ∈ 𝛑 𝜔 𝜃 , 𝝅 𝜔 𝜃 . Then the constraints become: 𝛑 𝜔 𝜃 ≤𝛑 𝜔 𝜃 ≤ 𝝅 𝜔 𝜃 ∀𝜔∈Ω 𝜔∈Ω 𝛑 𝜔 𝜃 =1

Robust spans the distance between Ex-Post and Bayesian Robust incentive compatibility and individual rationality contain ex- post and Bayesian IC and IR as special cases Suppose that the set 𝒫 𝝅 is a singleton, i.e. 𝒫 𝝅 ={ 𝝅 }. Then robust IR becomes ex-interim IR: 𝜔∈Ω 𝝅 𝜔 𝜃 𝑈(𝜃, 𝝅 ,𝜃,𝜔) ≥0 ∀ 𝜃∈Θ Instead suppose that the consistent set is such that for all 𝜃∈Θ and 𝜔∈Ω, 𝛑 𝜔 𝜃 ∈ 0,1 . Then the set of ex-post IR constraints appears in the robust IR set: 𝑈(𝜃, 𝝅 ,𝜃,𝜔)≥0 ∀ 𝜃∈Θ and 𝜔∈Ω Therefore, robust IR and IC spans the traditional set of constraints.

Is robust enough? Any reasonable estimation procedure will return a consistent set, 𝒫( 𝝅 ), as the entire set of possible distributions, 𝑃(Θ×Ω), i.e. 𝒫 𝝅 =𝑃(Θ×Ω). This is because the definition guarantees that the true distribution, 𝝅, is in the consistent set. We really want to say that the true distribution is in the consistent set with high probability: Let 𝑃(𝐴) be the set of probability distributions over the set 𝐴. Then the space of all probability distributions over Θ×Ω can be represented as 𝑃(Θ×Ω). A subset 𝒫 𝜖 ( 𝝅 )⊆𝑃(Θ×Ω) is a 𝜖- consistent set of distributions for the estimated distribution 𝝅 if the true distribution, 𝝅, is in 𝒫 𝜖 ( 𝝅 ) with probability 1−𝜖 and 𝝅 ∈ 𝒫 𝜖 ( 𝝅 ).

Finding the Consistent Set This is all well and good, but how do we actually find the consistent set if we just observe a bunch of samples from the distribution? The best way that I’ve found is to use Bayesian techniques A bivariate discrete distribution can generally be modeled as a categorical distribution (every possible combination of reports is a category) You can learn a categorical distribution using a Dirichlet distribution (it is the conjugate prior of the categorical distribution) Start with a Dirichlet distribution with an uninformative prior (𝜶=𝟏), then increment the element of 𝜶 by 1 every time that element is seen, i.e. if you see sample (𝜃′,𝜔′) then 𝜶 𝜃 ′ , 𝜔 ′ =𝜶 𝜃 ′ , 𝜔 ′ +1 The posterior is another Dirichlet distribution parameterized by the new 𝜶 Once the posterior Dirichlet distribution is calculated, you can sample from the distribution to get empirical confidence intervals Note that this can be done at the level of conditional distributions instead of the full distribution.

What happens in practice? True distribution is a discretized bivariate normal distribution Sample from true distribution N times Estimate the distribution using the previously described Bayesian procedure Calculate empirical confidence intervals for each element of the conditional distribution as an interval with odds of being outside of the interval as 𝜖 Ω , i.e. 𝛑 𝜔 𝜃 ∈ 𝛑 𝜔 𝜃 , 𝝅 𝜔 𝜃 with probability 1− 𝜖 Ω . This is because we are need the entire distribution to be inside of the interval Parameters unless otherwise specified Correlation = .5 𝜖=.05 Θ={1,2,…,10}, 𝑣 𝜃 =𝜃 Ω={1,2,…,10}

Open questions in robust mechanism design How do we optimally compute the 𝜖-consistent set? Can we bound the potential loss due to the lock of guaranteed IC and IR for 𝜖-robust mechanisms? If we can bound the maximum payment, this should be possible Can we use a prior over beliefs (the Dirichlet distribution we compute, for example) to construct a mechanism that depends on the reported belief of the bidder? What is the optimal approach to binning? What is the sample complexity of an 𝜖-approximation to the optimal mechanism? This will depend on characteristics of the underlying distribution (Albert, Conitzer, and Stone 2017 (AAMAS)) Can we characterize the sample complexity in terms of the separation between points in belief space? Are there simpler robust mechanisms that scale more easily to more bidders?

How do we learn online? Everything we’ve discussed assumes that someone hands us a bunch of samples. In reality, we will need to elicit these reports from bidders in multiple rounds, but this is not an easy task If bidder’s are myopic (i.e., they don’t lie to mislead future auctions) Then if the mechanism fails to be IR, we won’t see a report for that bidder type at all, skewing the distribution If the mechanism fails to be IC, we will see a wrong report, skewing the distribution Can use an ex-post mechanism for early rounds, and then stop learning. What is the regret? If bidders are strategic They may lie even when the mechanism is ex-post in order to influence future rounds Must give strict incentives to tell the truth in early rounds