Peer Prediction conitzer@cs.duke.edu.

Slides:



Advertisements
Similar presentations
Eliciting Honest Feedback 1.Eliciting Honest Feedback: The Peer-Prediction Model (Miller, Resnick, Zeckhauser) 2.Minimum Payments that Reward Honest Feedback.
Advertisements

Mixed Strategies.
CPS Bayesian games and their use in auctions Vincent Conitzer
Pondering more Problems. Enriching the Alice-Bob story Go to AGo to B Go to A Alice Go to B Go to A Go to B Go shoot pool Alice.
Non myopic strategy Truth or Lie?. Scoring Rules One important feature of market scoring rules is that they are myopic strategy proof. That means that.
4 Why Should we Believe Politicians? Lupia and McCubbins – The Democratic Dilemma GV917.
Game Theory, Mechanism Design, Differential Privacy (and you). Aaron Roth DIMACS Workshop on Differential Privacy October 24.
Cognitive Biases 2 Incomplete and Unrepresentative Data.
Adverse Selection Asymmetric information is feature of many markets
An Introduction to Game Theory Part IV: Games with Imperfect Information Bernhard Nebel.
Chapter 9 Games with Imperfect Information Bayesian Games.
Mechanisms for Making Crowds Truthful Andrew Mao, Sergiy Nesterko.
What Kind of Student are You Self Discovery. Before You Start  This is not a test, this is a chance to tell the truth about what kind of student you.
Derivative Action Learning in Games Review of: J. Shamma and G. Arslan, “Dynamic Fictitious Play, Dynamic Gradient Play, and Distributed Convergence to.
A Truth Serum for Sharing Rewards Arthur Carvalho Kate Larson.
Table of Contents Logarithm Properties - Product Rule The Product Rule for logarithms states that... read as “the log of the product is the sum of the.
3.1.4 Types of Games. Strategic Behavior in Business and Econ Outline 3.1. What is a Game ? The elements of a Game The Rules of the Game:
ECO290E: Game Theory Lecture 13 Dynamic Games of Incomplete Information.
Simple examples of the Bayesian approach For proportions and means.
Great Theoretical Ideas in Computer Science.
Auctions serve the dual purpose of eliciting preferences and allocating resources between competing uses. A less fundamental but more practical reason.
CPS Proper scoring rules Vincent Conitzer
How I Should Talk to My Teacher
Bayesian networks Chapter 14 Slide Set 2. Constructing Bayesian networks 1. Choose an ordering of variables X 1, …,X n 2. For i = 1 to n –add X i to the.
Midterm Scores Total of 200 points, 40 per question. A B— C— D—50-79 F
Incomplete Information and Bayes-Nash Equilibrium.
Naïve Bayes Classification Recitation, 1/25/07 Jonathan Huang.
1 Neural Codes. 2 Neuronal Codes – Action potentials as the elementary units voltage clamp from a brain cell of a fly.
1 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A AAA A AA A.
Creating Institutions to Address Externalities TMS.
Final Lecture.
Midterm Scores Total of 200 points, 40 per question. A B— C— D—50-79 F
Correlated equilibria, good and bad: an experimental study
Lecture 7: Constrained Conditional Models
A Conceptual Design of Multi-Agent based Personalized Quiz Game
Step 1: Specify a null hypothesis
Bayesian games and mechanism design
Market Scoring Rules
Bayesian games and their use in auctions
Comp/Math 553: Algorithmic Game Theory Lecture 08
Propositional Logic.
Unit 5: Hypothesis Testing
The Beta Reputation System
Bidding on an Antique.
Information Design: A unified Perspective Prior information
Roberto Battiti, Mauro Brunato
CSE 105 theory of computation
Intro to Information Design (and some Basic Bayesian Persuasion)
To Detect and Correct: Norm Violations and Their Enforcement
Instructors: Fei Fang (This Lecture) and Dave Touretzky
CS 188: Artificial Intelligence Fall 2007
Chapter 29 Game Theory Key Concept: Nash equilibrium and Subgame Perfect Nash equilibrium (SPNE)
Vincent Conitzer Mechanism design Vincent Conitzer
Vincent Conitzer CPS 173 Mechanism design Vincent Conitzer
Games with Imperfect Information Bayesian Games
I can work with different people in my class
TECHNIQUES OF INTEGRATION
Computing probabilities using Expect problem-solving Trees: A worked example Jim Blythe USC/ISI.
Learning outcomes Knowledge Skills
CS 416 Artificial Intelligence
1.8 Safe and happy online I can discuss my online life.
Game Theory: The Nash Equilibrium
CSE 105 theory of computation
Y7 Assessment – Calculations with Fractions
Safe and happy online I can discuss my online life.
Instructor: Vincent Conitzer
Bayes Nash Implementation
Truth tables.
CPS Bayesian games and their use in auctions
Non-Myopic Strategic Bluffing
Presentation transcript:

Peer Prediction conitzer@cs.duke.edu

Example setup We are evaluating a theme park which can be either Good or Bad P(G) = .8 If you visit, you can have an Enjoyable or an Unpleasant experience P(E|G) = .9, P(E|B) = .7 We ask people to report their experiences and want to reward them for accurate reporting I had fun. E quality: good The problem: we will never find out the true quality / experience. Another nice application: peer grading (of, say, essays) in a MOOC.

Solution: use multiple raters I had fun. E quality: good I had fun. E’ Rough idea: other agent likely (though not surely) had a similar experience Evaluate a rater by how well her report matches the other agent’s report How might this basic idea fail?

Simple approach: output agreement Receive 1 if you agree, 0 otherwise What’s the problem? What is P(other reports E | I experienced U) (given that the other reports truthfully)? P(E’|U) = P(U and E’) / P(U) P(U and E’) = P(U, E’, G) + P(U, E’, B) = .8 .1 .9 + .2 .3 .7 = .072+.042 = .114 P(U) = P(U,G) + P(U,B) = .8 .1 + .2 .3 = .08 + .06 = .14 So P(E’|U) = .114 / .14 = .814 P(E’|E) = P(E and E’) / P(E) P(E and E’) = P(E, E’, G) + P(E, E’, B) = .8 .9 .9 + .2 .7 .7 = .648 + .098 = .746 P(E) = P(E,G) + P(E,B) = .8 .9 + .2 .7 = .72 + .14 = .86 So P(E’|E) = .746 / .86 = .867

The “1/Prior” mechanism [Jurca&Faltings’08] Receive 1/P(s) if you agree on signal s, 0 otherwise P(E) = .86 and P(U) = .14 so 1/P(E)=1.163 and 1/P(U)=7.143 P(E’|U) (1/P(E’)) = .814*1.163 =.95 … but, P(U’|U)(1/P(U’)) = .186*7.143=1.33 Why does this work? (When does this work?) Need, for all signals s, t: P(s’|s)/P(s’) > P(t’|s)/P(t’) Equivalently, for all signals s, t: P(s,s’)/P(s’) > P(s,t’)/P(t’) Equivalently, for all signals s, t: P(s|s’) > P(s|t’)

An example where the “1/Prior” mechanism does not work P(A|Good)=.9, P(B|Good)=.1, P(C|Good)=0 P(A|Bad)=.4, P(B|Bad)=.5, P(C|Bad)=.1 P(Good)=P(Bad)=.5 Note that P(B|B’) < P(B|C’), so the condition from the previous slide is violated Suppose I saw B and the other player reports honestly P(B’|B) = P(B’, Good|B) + P(B’, Bad|B) = P(B’|Good)P(Good|B) + P(B’|Bad)P(Bad|B) = .1*(1/6) + .5*(5/6) = 13/30 P(B’) = 3/10, so expected reward for reporting B is 130/90 = 13/9 = 1.44 P(C’|B) = P(C’, Good|B) + P(C’, Bad|B) = P(C’|Good)P(Good|B) + P(C’|Bad)P(Bad|B) = 0*(1/6) + .1*(5/6) = 1/12 P(C’) = 1/20, so expected reward for reporting C is 20/12 = 5/3 = 1.67

Better idea: use proper scoring rules Assuming the other reports truthfully, can infer a conditional distribution over the other’s report given my report Reward me according to a proper scoring rule! Suppose we use the logarithmic rule Reporting E  predicting the other reports E’ with P(E’|E) = .867 Reporting U  predicting the other reports E’ with P(E’|U) = .814 E.g., if report E and the other reports U’, I get ln(P(U’|E)) = ln .133 In what sense does this work? Truthful reporting is an equilibrium

… as a Bayesian game A player’s type (private information): experience the player truly had (E or U) Note types are correlated (only displaying player 1’s payoffs) E’ U’ true experiences: E and U’ (prob. .114) E’ U’ true experiences: E and E’ (prob. .746) E ln .867 ln .133 ln .814 ln .186 E ln .867 ln .133 ln .814 ln .186 U U E’ U’ true experiences: U and U’ (prob. .026) E’ U’ true experiences: U and E’ (prob. .114) E ln .867 ln .133 ln .814 ln .186 E ln .867 ln .133 ln .814 ln .186 U U

true experiences: E and E’ (prob. .746) E’ U’ true experiences: E and U’ (prob. .114) E’ U’ E -.143 -2.017 -.205 -1.682 E -.143 -2.017 -.205 -1.682 U U true experiences: U and E’ (prob. .114) E’ U’ E’ U’ true experiences: U and U’ (prob. .026) E -.143 -2.017 -.205 -1.682 E -.143 -2.017 -.205 -1.682 U U observe E: report E observe U: report U observe E: report E observe U: report E observe E: report U observe U: report U observe E: report U observe U: report E observe E: report E observe U: report U -.404, -.404 -.152, -.405 -1.970, -.412 -1.718, -.413 -.405, -.152 -.143, -.143 -2.017, -.205 -1.755, -.196 -.412, -1.970 -.205, -2.017 -1.682, -1.682 -1.475, -1.729 -.413, -1.718 -.196, -1.755 -1.729, -1.475 -1.512, -1.512 observe E: report E observe U: report E observe E: report U observe U: report U observe E: report U observe U: report E

Downsides (and how to fix them, maybe?) Multiplicity of equilibria Completely uninformative equilibria Uselessly informative equilibria: Users may be supposed to evaluate whether the image contains a person, but instead reach an equilibrium where they evaluate whether the top-left pixel is blue Need to know the prior distribution beforehand Explicitly report beliefs as well [Prelec’04] Bonus-penalty mechanism [Dasgupta&Ghosh’13, Shnayder et al.’16]: Suppose there are 3 tasks (e.g., 3 essays to grade) You get a bonus for agreeing on the third task Agents don’t know how the tasks are ordered You get a penalty for agent 1’s report on 1 agreeing with agent 2’s report on 2 Use a limited number of trusted reports (e.g., the instructor grades) …?