Incentive Compatible Regression Learning Ofer Dekel, Felix A. Fischer and Ariel D. Procaccia.

Slides:



Advertisements
Similar presentations
Truthful Mechanisms for Combinatorial Auctions with Subadditive Bidders Speaker: Shahar Dobzinski Based on joint works with Noam Nisan & Michael Schapira.
Advertisements

Real-Time Competitive Environments: Truthful Mechanisms for Allocating a Single Processor to Sporadic Tasks Anwar Mohammadi, Nathan Fisher, and Daniel.
Statistical Machine Learning- The Basic Approach and Current Research Challenges Shai Ben-David CS497 February, 2007.
6.896: Topics in Algorithmic Game Theory Lecture 20 Yang Cai.
Blackbox Reductions from Mechanisms to Algorithms.
1 AI and Economics: The Dynamic Duo Ariel Procaccia Center for Research on Computation and Society Harvard SEAS AI AND ECONOMICS DYNAMIC DUO THE.
Mechanism Design, Machine Learning, and Pricing Problems Maria-Florina Balcan.
Distributional Property Estimation Past, Present, and Future Gregory Valiant (Joint work w. Paul Valiant)
A sublinear Time Approximation Scheme for Clustering in Metric Spaces Author: Piotr Indyk IEEE FOCS 1999.
Learning Voting Trees Ariel D. Procaccia, Aviv Zohar, Yoni Peleg, Jeffrey S. Rosenschein.
6.896: Topics in Algorithmic Game Theory Lecture 11 Constantinos Daskalakis.
1 Regret-based Incremental Partial Revelation Mechanism Design Nathanaël Hyafil, Craig Boutilier AAAI 2006 Department of Computer Science University of.
1 Learning with continuous experts using Drifting Games work with Robert E. Schapire Princeton University work with Robert E. Schapire Princeton University.
Seminar In Game Theory Algorithms, TAU, Agenda  Introduction  Computational Complexity  Incentive Compatible Mechanism  LP Relaxation & Walrasian.
Machine Learning Week 2 Lecture 1.
1 Truthful Mechanism for Facility Allocation: A Characterization and Improvement of Approximation Ratio Pinyan Lu, MSR Asia Yajun Wang, MSR Asia Yuan Zhou,
Nonstochastic Multi-Armed Bandits With Graph-Structured Feedback Noga Alon, TAU Nicolo Cesa-Bianchi, Milan Claudio Gentile, Insubria Shie Mannor, Technion.
The loss function, the normal equation,
On the Limits of Dictatorial Classification Reshef Meir School of Computer Science and Engineering, Hebrew University Joint work with Shaull Almagor, Assaf.
Sum of Us: Strategyproof Selection From the Selectors Noga Alon, Felix Fischer, Ariel Procaccia, Moshe Tennenholtz 1.
Reshef Meir, Ariel D. Procaccia, and Jeffrey S. Rosenschein.
Ariel D. Procaccia (Microsoft)  Best advisor award goes to...  Thesis is about computational social choice Approximation Learning Manipulation BEST.
MIX AND MATCH Itai Ashlagi, Felix Fischer, Ian Kash, Ariel Procaccia (Harvard SEAS)
Machine Learning Week 2 Lecture 2.
Strategy-Proof Classification Reshef Meir School of Computer Science and Engineering, Hebrew University A joint work with Ariel. D. Procaccia and Jeffrey.
Limitations of VCG-Based Mechanisms Shahar Dobzinski Joint work with Noam Nisan.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
1 A Confidence Interval for the Misclassification Rate S.A. Murphy & E.B. Laber.
Probably Approximately Correct Model (PAC)
Vapnik-Chervonenkis Dimension
SECOND PART: Algorithmic Mechanism Design. Mechanism Design MD is a subfield of economic theory It has a engineering perspective Designs economic mechanisms.
Learning From Data Chichang Jou Tamkang University.
Junta Distributions and the Average Case Complexity of Manipulating Elections A. D. Procaccia & J. S. Rosenschein.
Learning to Identify Winning Coalitions in the PAC Model A. D. Procaccia & J. S. Rosenschein.
Gossip-Based Aggregation of Trust in Decentralized Reputation Systems Ariel D. Procaccia, Yoram Bachrach, and Jeffrey S. Rosenschein.
Visual Recognition Tutorial
Competitive Analysis of Incentive Compatible On-Line Auctions Ron Lavi and Noam Nisan SISL/IST, Cal-Tech Hebrew University.
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 13 June 22, 2005
Strategy-Proof Classification Reshef Meir School of Computer Science and Engineering, Hebrew University A joint work with Ariel. D. Procaccia and Jeffrey.
Bayesian and non-Bayesian Learning in Games Ehud Lehrer Tel Aviv University, School of Mathematical Sciences Including joint works with: Ehud Kalai, Rann.
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 21.
Benk Erika Kelemen Zsolt
Market Design and Analysis Lecture 5 Lecturer: Ning Chen ( 陈宁 )
Overview of Supervised Learning Overview of Supervised Learning2 Outline Linear Regression and Nearest Neighbors method Statistical Decision.
Online Passive-Aggressive Algorithms Shai Shalev-Shwartz joint work with Koby Crammer, Ofer Dekel & Yoram Singer The Hebrew University Jerusalem, Israel.
Maria-Florina Balcan Mechanism Design, Machine Learning and Pricing Problems Maria-Florina Balcan Joint work with Avrim Blum, Jason Hartline, and Yishay.
Topics in Algorithms 2007 Ramesh Hariharan. Tree Embeddings.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 22.
Secret Sharing Non-Shannon Information Inequalities Presented in: Theory of Cryptography Conference (TCC) 2009 Published in: IEEE Transactions on Information.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
5.3 Algorithmic Stability Bounds Summarized by: Sang Kyun Lee.
Robert Pless, CS 546: Computational Geometry Lecture #3 Last time Several convex hull algorithms. Lower bound of O(n log n) –O(n log h) for output sensitive.
Debrup Chakraborty Non Parametric Methods Pattern Recognition and Machine Learning.
Bayesian Algorithmic Mechanism Design Jason Hartline Northwestern University Brendan Lucier University of Toronto.
PAC-Bayesian Analysis of Unsupervised Learning Yevgeny Seldin Joint work with Naftali Tishby.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Deep Feedforward Networks
Stochastic Streams: Sample Complexity vs. Space Complexity
Empirical risk minimization
Understanding Generalization in Adaptive Data Analysis
Privacy-Preserving Classification
CSCI B609: “Foundations of Data Science”
CSCI B609: “Foundations of Data Science”
The loss function, the normal equation,
Approximation and Generalization in Neural Networks
Empirical risk minimization
Machine learning overview
Generalization bounds for uniformly stable algorithms
Presentation transcript:

Incentive Compatible Regression Learning Ofer Dekel, Felix A. Fischer and Ariel D. Procaccia

Lecture Outline Until now: applications of learning to game theory. Now: merge. The model: –Motivation –The learning game Three levels of generality: –Distributions which are degenerate at one point –Uniform distributions –The general setting ModelDegenerateUniformGeneral

Motivation Internet search company: improve performance by learning ranking function from examples. Ranking function assigns real value to every (query,answer). Employ experts to evaluate examples. Different experts may have diff. interests and diff. ideas of good output. Conflict  Manipulation  Bias in training set. ModelDegenerateUniformGeneral

Jaguar vs. Panthera Onca (“Jaguar”, jaguar.com) ModelDegenerateUniformGeneral

Regression Learning Input space X=R k ((query,answer) pairs). Function class F:X  R (ranking functions). Target function o:X  R. Distribution  over X. Loss function l (a,b). –Abs. loss: l (a,b)=|a-b|. –Squared loss: l (a,b)=(a-b) 2. Learning process: –Given: Training set S={(x i,o(x i ))}, i=1,...,m, x i sampled from . –R(h)=E x  [ l (h(x),o(x))]. –Find: h  F to minimize R(h). ModelDegenerateUniformGeneral

Our Setting Input space X=R k ((query,answer) pairs). Function class F (ranking functions). Set of players N={1,...,n} (experts). Target functions o i :X  R. Distributions  i over X. Training set? ModelDegenerateUniformGeneral

The Learning Game  i: controls x ij, j=1,...,m, sampled w.r.t.  i (common knowledge). Private info of i: o i (x ij )=y ij, j=1,...,m. Strategies of i: y’ ij, j=1,...,m. h is obtained by learning S={(x ij,y’ ij )} Cost of i: R i (h)=E x  i [ l (h(x),o i (x))]. Goal: Social Welfare (please avg. player). ModelDegenerateUniformGeneral

Example: The learning game with ERM Parameters: X=R, F=Constant Functions, l (a,b)=|a-b|, N={1,2}, o 1 (x)=1, o 2 (x)=2,  1 =  2 =uniform dist on [0,1000]. Learning algorithm: Empirical Risk Minimization (ERM) –Minimize R’(h,S)=1/|S|   (x,y)  S l (h(x),y). 1 2 ModelDegenerateUniformGeneral

Degenerate Distributions: ERM with abs. loss The Game: –Players: N={1,...n} –  i : degenerate at x i. –  i: controls x i. –Private info of i: o i (x i )=y i. –Strategies of i: y’ i. –Cost of i: R i (h)= l (h(x i ),y i ). Theorem: If l = absolute loss and F is convex. Then ERM is group incentive compatible. ModelDegenerateUniformGeneral

ERM with superlinear loss Theorem: l is “superlinear”, F is convex, |F|  2, F is not “full” on x 1,...,x n. Then  y 1,...,y n such that there is incentive to lie. Example: X=R, F=Constant Functions, l (a,b)=(a-b) 2, N={1,2}. ModelDegenerateUniformGeneral

Uniform dist. over samples The Game: –Players: N={1,...n} –  i : Discrete uniform on {x i1,...,x im } –  i: controls x ij, j=1,...,m –Private info of i: o i (x ij )=y ij. –Strategies of i: y’ ij, j=1,...,m. –Cost of i: R i (h)= R’ i (h,S)= 1/m  j l (h(x ij ),y ij ). ModelDegenerateUniformGeneral

ERM with abs. loss is not IC ModelDegenerateUniformGeneral 1 0

VCG to the Rescue Use ERM. Each player pays  j  i R’ j (h,S). Each player’s total cost is R’ i (h,S)+  j  i R j ’(h,S) =  j R’ j (h,S). Truthful for any loss function. VCG has many faults: –Not group incentive compatible. –Payments problematic in practice. Would like (group) IC mechanisms w/o payments. ModelDegenerateUniformGeneral

Mechanisms w/o Payments Absolute loss.  -approximation mechanism: gives an  - approximation of the social welfare. Theorem (upper bound): There exists a group IC 3-approx mechanism for constant functions over R k and homogeneous linear functions over R. Theorem (lower bound): There is no IC (3-  )- approx mechanism for constant/hom. lin. functions over R k. Conjecture: There is no IC mechanism with bounded approx. ratio for hom. lin. functions over R k, k  2. ModelDegenerateUniformGeneral

Proof of Lower Bound k-1 k k k k k ModelDegenerateUniformGeneral

Proof of Lower Bound k k-1 k k k ModelDegenerateUniformGeneral

Generalization Theorem: If  f, –(1)  i, |R’ i (f,S)-R i (f)|   /2 –(2) |R’(f,S)-1/n  i R i (f)|   /2 Then: –(Group) IC in uniform   -(group) IC in general. –  -approx in uniform   -approx up to additive  in general. If F has bounded complexity, m=  (log(1/  )/  ), then cond. (1) holds with prob. 1- . Cond. (2) is obtained if (1) occurs for all i. Taking  /n adds factor of logn. ModelDegenerateUniformGeneral

Discussion Given m large enough, with prob. 1-  VCG is  -truthful. This holds for any loss function. Given m large enough, abs loss,  mechanism w/o payments which is  -group IC and 3- approx for constant functions and hom. lin. functions. Most important direction for future work: extending to other models of learning, such as classification. ModelDegenerateUniformGeneral