Statistical Cost Sharing: Learning Fair Cost Allocations from Samples

Slides:



Advertisements
Similar presentations
Sublinear-time Algorithms for Machine Learning Ken Clarkson Elad Hazan David Woodruff IBM Almaden Technion IBM Almaden.
Advertisements

On allocations that maximize fairness Uriel Feige Microsoft Research and Weizmann Institute.
A Short Tutorial on Cooperative Games
CS4432: Database Systems II
Fast Algorithms For Hierarchical Range Histogram Constructions
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
Learning Cooperative Games Maria-Florina Balcan, Ariel D. Procaccia and Yair Zick (to appear in IJCAI 2015)
Support Vector Machines (SVMs) Chapter 5 (Duda et al.)
The Communication Complexity of Coalition Formation Among Autonomous Agents A. D. Procaccia & J. S. Rosenschein.
Northwestern University Winter 2007 Machine Learning EECS Machine Learning Lecture 13: Computational Learning Theory.
Vapnik-Chervonenkis Dimension
1 Sampling Lower Bounds via Information Theory Ziv Bar-Yossef IBM Almaden.
CISS Princeton, March Optimization via Communication Networks Matthew Andrews Alcatel-Lucent Bell Labs.
Learning to Identify Winning Coalitions in the PAC Model A. D. Procaccia & J. S. Rosenschein.
Preference Analysis Joachim Giesen and Eva Schuberth May 24, 2006.
SVM Support Vectors Machines
Maria-Florina Balcan A Theoretical Model for Learning from Labeled and Unlabeled Data Maria-Florina Balcan & Avrim Blum Carnegie Mellon University, Computer.
Computational aspects of stability in weighted voting games Edith Elkind (NTU, Singapore) Based on joint work with Leslie Ann Goldberg, Paul W. Goldberg,
Topic Models in Text Processing IR Group Meeting Presented by Qiaozhu Mei.
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
1 - CS7701 – Fall 2004 Review of: Detecting Network Intrusions via Sampling: A Game Theoretic Approach Paper by: – Murali Kodialam (Bell Labs) – T.V. Lakshman.
Joel Oren, University of Toronto
Lecture 5 Introduction to Sampling Distributions.
CS 8751 ML & KDDComputational Learning Theory1 Notions of interest: efficiency, accuracy, complexity Probably, Approximately Correct (PAC) Learning Agnostic.
Pseudo-random generators Talk for Amnon ’ s seminar.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Support Vector Machines (SVMs) Chapter 5 (Duda et al.) CS479/679 Pattern Recognition Dr. George Bebis.
1 CS 391L: Machine Learning: Computational Learning Theory Raymond J. Mooney University of Texas at Austin.
Approximation algorithms for combinatorial allocation problems
Learning to Align: a Statistical Approach
CS 9633 Machine Learning Support Vector Machines
Estimating the Value of a Parameter
Estimating the Value of a Parameter Using Confidence Intervals
University of Texas at El Paso
Chapter 7. Classification and Prediction
Stochastic Streams: Sample Complexity vs. Space Complexity
Vitaly Feldman and Jan Vondrâk IBM Research - Almaden
Computational Learning Theory
Moran Feldman The Open University of Israel
Understanding Generalization in Adaptive Data Analysis
Algorithmic Game Theory and Internet Computing
CPS Cooperative/coalitional game theory
Dana Moshkovitz The Institute For Advanced Study
LSI, SVD and Data Management
INTRODUCTION TO Machine Learning
Vitaly (the West Coast) Feldman
Data Integration with Dependent Sources
Random Sampling over Joins Revisited
Statistics for the Social Sciences
Mechanism Design via Machine Learning
OUTLINE Lecture 5 A. Review of Lecture 4 B. Special SLR Models
Computational Learning Theory
Summarizing Data by Statistics
Estimating the Value of a Parameter
LEARNING Chapter 18b and Chuck Dyer
Computational Learning Theory
Feifei Li, Ching Chang, George Kollios, Azer Bestavros
Welcome to the Kernel-Club
Katsushige FUJIMOTO Fukushima University
Computational Learning Theory Eric Xing Lecture 5, August 13, 2010
Selfish Load Balancing
Topic Models in Text Processing
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Network-Wide Routing Oblivious Heavy Hitters
2.3. Measures of Dispersion (Variation):
Classifier-Based Approximate Policy Iteration
INTRODUCTION TO Machine Learning 3rd Edition
STATISTICS INFORMED DECISIONS USING DATA
Presentation transcript:

Statistical Cost Sharing: Learning Fair Cost Allocations from Samples Eric Balkanski Harvard University Joint work with Umar Syed and Sergei Vassilvitskii (Google NYC) Dagstuhl Seminar on Game Theory in AI, Logic, and Algorithms \max_S\ & f(S) \\ \notag \text{s.t. } &\sum_{i \in S} p_i \leq B, \\ \notag & p_i \geq c_i \text{ for all } i \in S. 3/16/17 Dagstuhl Seminar

Motivating Example: Attributing Battery Consumption Data: (apps running, units of battery usage) ( , 8) ( , 3) ( , 9) Goal: fair blaming : 40%, : 40%, : 20% 3/16/17 Dagstuhl Seminar

Cost Sharing Aims to find an equitable way to split the cost of a service among players N Cost function Cost allocation Multiple solutions suggested: Shapley value, core, nucleolus… But: Assumes complete knowledge of cost function C In previous example, only given samples of C 3/16/17 Dagstuhl Seminar

Statistical Cost Sharing Definition: Given samples (S1, C(S1)), … (Sm, C(Sm)) where Si drawn i.i.d. from a distribution D, the Statistical Cost Sharing problem asks to find a cost allocation . [Balcan, Procaccia, and Zick 15]: “It is the authors’ opinion that the information required in order to compute cooperative solution concepts (much more than the computation complexity) is a major obstacle to their widespread implementation” 3/16/17 Dagstuhl Seminar

Outline The core Previously and concurrently studied [Balcan et al. 15, 16] The Shapley value Never studied before in the statistical cost sharing context 3/16/17 Dagstuhl Seminar

The core 3/16/17 Dagstuhl Seminar

The Core “no group of players has an incentive to deviate” Balance: must allocate the total cost: Core [Gillies 59]: is in the core of C if for all Definition [Balcan et al. 15]: Given poly(n, 1/δ) samples from D, is in the probably stable core if 3/16/17 Dagstuhl Seminar

Prior Approach C’ (learned) cost function C data S1, C(S1) … Sm, C(Sm) Traditional Cost Sharing Learning (PAC) data S1, C(S1) … Sm, C(Sm) core Statistical Cost Sharing Thm [Balcan et al. 15]: Assume C’ PAC-learns C from samples, then a cost allocation that is in the core of C’ is in the probably stable core of C. 3/16/17 Dagstuhl Seminar

Direct Approach cost function C data S1, C(S1) … Sm, C(Sm) core Thm [B., Syed, Vassilvitskii 16, Balcan et al. 16]: If C has a non empty core, then C has a computable probably stable core . Traditional Cost Sharing Learning (PAC) data S1, C(S1) … Sm, C(Sm) core Statistical Cost Sharing 3/16/17 Dagstuhl Seminar

High Level Overview of Analysis Compute any vector that satisfies the core property on samples, i.e., for all samples S. Bound the generalization error to another set drawn from D using tools from theoretical machine learning. Using VC dimension: sample complexity linear in n [B., Syed, Vassilvitskii 16]. Using Rademacher complexity: sample complexity logarithmic in n, but dependent on the spread of C and weaker relaxation of the core [B., Syed, Vassilvitskii 16]. 3/16/17 Dagstuhl Seminar

The Shapley value 3/16/17 Dagstuhl Seminar

Shapley Value Unique cost allocation satisfying four natural axioms Definition: Given poly(n) samples from D, an algorithm α-approximates the Shapley value of C over D if it computes estimates s.t. for all i, 3/16/17 Dagstuhl Seminar

Impossibility for Approximating Shapley Value cost function C Thm: There is no constant α > 0 such that coverage functions have α-approximable Shapley value, even over the uniform distribution D. Traditional Cost Sharing Learning (PAC) data S1, C(S1) … Sm, C(Sm) Shapley value Statistical Cost Sharing Thm: Functions with curvature κ have √(1 – κ)-approximable Shapley value over the uniform distribution. Moreover, this is tight. 3/16/17 Dagstuhl Seminar

Data Dependent Shapley Value We define four novel natural axioms that are dependent on D There exists a unique cost allocation satisfying these four axioms These data dependent Shapley value is (1-ε)-approximable from samples for any distribution and any function. 3/16/17 Dagstuhl Seminar

Conclusion Studied cost sharing problem where the cost function is unknown and only given sampled data from this function Possible to compute cost allocations from samples that satisfy a simple relaxation of the core for all functions with a non-empty core Also possible to approximate arbitrarily well data-dependent Shapley value, which is the unique cost allocation satisfying four novel axioms. Future work: other cost sharing methods, other classes of functions, better sample complexity bounds,… 3/16/17 Dagstuhl Seminar