1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin and Andreas Krause.

Slides:



Advertisements
Similar presentations
Truthful Mechanisms for Combinatorial Auctions with Subadditive Bidders Speaker: Shahar Dobzinski Based on joint works with Noam Nisan & Michael Schapira.
Advertisements

On allocations that maximize fairness Uriel Feige Microsoft Research and Weizmann Institute.
1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Joint work with Andreas Krause 1 Daniel Golovin.
Submodular Set Function Maximization A Mini-Survey Chandra Chekuri Univ. of Illinois, Urbana-Champaign.
Submodularity for Distributed Sensing Problems Zeyn Saigol IR Lab, School of Computer Science University of Birmingham 6 th July 2010.
Submodular Set Function Maximization via the Multilinear Relaxation & Dependent Rounding Chandra Chekuri Univ. of Illinois, Urbana-Champaign.
Minimizing Seed Set for Viral Marketing Cheng Long & Raymond Chi-Wing Wong Presented by: Cheng Long 20-August-2011.
Spread of Influence through a Social Network Adapted from :
Cost-effective Outbreak Detection in Networks Jure Leskovec, Andreas Krause, Carlos Guestrin, Christos Faloutsos, Jeanne VanBriesen, Natalie Glance.
Randomized Sensing in Adversarial Environments Andreas Krause Joint work with Daniel Golovin and Alex Roper International Joint Conference on Artificial.
Study Group Randomized Algorithms 21 st June 03. Topics Covered Game Tree Evaluation –its expected run time is better than the worst- case complexity.
Parallel Double Greedy Submodular Maxmization Xinghao Pan, Stefanie Jegelka, Joseph Gonzalez, Joseph Bradley, Michael I. Jordan.
ANDREW MAO, STACY WONG Regrets and Kidneys. Intro to Online Stochastic Optimization Data revealed over time Distribution of future events is known Under.
Maximizing the Spread of Influence through a Social Network
Online Distributed Sensor Selection Daniel Golovin, Matthew Faulkner, Andreas Krause theory and practice collide 1.
Submodular Dictionary Selection for Sparse Representation Volkan Cevher Laboratory for Information and Inference Systems - LIONS.
Near-Optimal Sensor Placements in Gaussian Processes Carlos Guestrin Andreas KrauseAjit Singh Carnegie Mellon University.
Efficient Informative Sensing using Multiple Robots
Optimal Marketing Strategies over Social Networks Jason Hartline (Northwestern), Vahab Mirrokni (Microsoft Research) Mukund Sundararajan (Stanford)
Near-optimal Nonmyopic Value of Information in Graphical Models Andreas Krause, Carlos Guestrin Computer Science Department Carnegie Mellon University.
Sensor placement applications Monitoring of spatial phenomena Temperature Precipitation... Active learning, Experiment design Precipitation data from Pacific.
1 Dynamic Resource Allocation in Conservation Planning 1 Daniel GolovinAndreas Krause Beth Gardner Sarah Converse Steve Morey.
1 Efficient planning of informative paths for multiple robots Amarjeet Singh *, Andreas Krause +, Carlos Guestrin +, William J. Kaiser *, Maxim Batalin.
Group Strategyproofness and No Subsidy via LP-Duality By Kamal Jain and Vijay V. Vazirani.
On Stochastic Minimum Spanning Trees Kedar Dhamdhere Computer Science Department Joint work with: Mohit Singh, R. Ravi (IPCO 05)
Efficiently handling discrete structure in machine learning Stefanie Jegelka MADALGO summer school.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
Active Learning for Probabilistic Models Lee Wee Sun Department of Computer Science National University of Singapore LARC-IMS Workshop.
Machine Learning and Optimization For Traffic and Emergency Resource Management. Milos Hauskrecht Department of Computer Science University of Pittsburgh.
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
Approximation algorithms for sequential testing of Boolean functions Lisa Hellerstein Polytechnic Institute of NYU Joint work with Devorah Kletenik (Polytechnic.
Approximating Minimum Bounded Degree Spanning Tree (MBDST) Mohit Singh and Lap Chi Lau “Approximating Minimum Bounded DegreeApproximating Minimum Bounded.
Chapter 10. Sampling Strategy for Building Decision Trees from Very Large Databases Comprising Many Continuous Attributes Jean-Hugues Chauchat and Ricco.
Randomized Composable Core-sets for Submodular Maximization Morteza Zadimoghaddam and Vahab Mirrokni Google Research New York.
Energy-Efficient Monitoring of Extreme Values in Sensor Networks Loo, Kin Kong 10 May, 2007.
Maximizing the Spread of Influence through a Social Network Authors: David Kempe, Jon Kleinberg, É va Tardos KDD 2003.
5 Maximizing submodular functions Minimizing convex functions: Polynomial time solvable! Minimizing submodular functions: Polynomial time solvable!
© 2009 Ilya O. Ryzhov 1 © 2008 Warren B. Powell 1. Optimal Learning On A Graph INFORMS Annual Meeting October 11, 2009 Ilya O. Ryzhov Warren Powell Princeton.
Adaptive Dependent Context BGL: Budgeted Generative (-) Learning Given nothing about training instances, pay for any feature [no “labels”, no “attributes”
A Unified Continuous Greedy Algorithm for Submodular Maximization Moran Feldman Roy SchwartzJoseph (Seffi) Naor Technion – Israel Institute of Technology.
Maximization Problems with Submodular Objective Functions Moran Feldman Publication List Improved Approximations for k-Exchange Systems. Moran Feldman,
Submodular Set Function Maximization A Mini-Survey Chandra Chekuri Univ. of Illinois, Urbana-Champaign.
Deterministic Algorithms for Submodular Maximization Problems Moran Feldman The Open University of Israel Joint work with Niv Buchbinder.
Aspects of Submodular Maximization Subject to a Matroid Constraint Moran Feldman Based on A Unified Continuous Greedy Algorithm for Submodular Maximization.
Maximizing Symmetric Submodular Functions Moran Feldman EPFL.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
Polyhedral Optimization Lecture 5 – Part 3 M. Pawan Kumar Slides available online
Unconstrained Submodular Maximization Moran Feldman The Open University of Israel Based On Maximizing Non-monotone Submodular Functions. Uriel Feige, Vahab.
Approximation algorithms for combinatorial allocation problems
Independent Cascade Model and Linear Threshold Model
Monitoring rivers and lakes [IJCAI ‘07]
Near-optimal Observation Selection using Submodular Functions
Vitaly Feldman and Jan Vondrâk IBM Research - Almaden
Moran Feldman The Open University of Israel
Joseph E. Gonzalez Postdoc, UC Berkeley AMPLab
Maximum Matching in the Online Batch-Arrival Model
Combinatorial Optimization Under Uncertainty
The Price of information in combinatorial optimization
Distributed Submodular Maximization in Massive Datasets
Independent Cascade Model and Linear Threshold Model
Framework for the Secretary Problem on the Intersection of Matroids
Coverage Approximation Algorithms
Cost-effective Outbreak Detection in Networks
Rule Selection as Submodular Function
Submodular Maximization Through the Lens of the Multilinear Relaxation
Submodular Maximization in the Big Data Era
Independent Cascade Model and Linear Threshold Model
Submodular Maximization with Cardinality Constraints
Guess Free Maximization of Submodular and Linear Sums
Presentation transcript:

1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin and Andreas Krause

2 Max K-Cover (Oil Spill Edition)

3 Submodularity Time Discrete diminishing returns property for set functions. ``Playing an action at an earlier stage only increases its marginal benefit''

4 The Greedy Algorithm Theorem [Nemhauser et al ‘78]

5 Stochastic Max K-Cover Asadpour & Saberi (`08): (1-1/e)-approx if sensors (independently) either work perfectly or fail completely. Bayesian: Known failure distribution. Adaptive: Deploy a sensor and see what you get. Repeat K times At 1 st location

6 Adaptive Submodularity Time Playing an action at an earlier stage only increases its marginal benefit expected (taken over its outcome) Gain more Gain less (i.e., at an ancestor) Select Item Stochastic Outcome

7 Adaptive Monotonicity

8 What is it good for? Allows us to generalize various results to the adaptive realm, including: (ln(n)+1)-approximation for Set Cover (1-1/e)-approximation for Max K-Cover, submodular maximization subject to a cardinality constraint

9 Recall the Greedy Algorithm Theorem [Nemhauser et al ‘78]

10 The Adaptive-Greedy Algorithm Theorem

11 [Adapt-monotonicity] - - ( ) - [Adapt-submodularity]

12 … The world-state dictates which path in the tree we’ll take. 1.For each node at layer i+1, 2.Sample path to layer j, 3.Play the resulting layer j action at layer i+1. How to play layer j at layer i+1 By adapt. submod., playing a layer earlier only increases it’s marginal benefit

13 [Adapt-monotonicity] - - ( ) - ( ) - [Def. of adapt-greedy] ( ) - [Adapt-submodularity]

14

Stochastic Max Cover is Adapt-Submod 1 3 Gain more Gain less adapt-greedy is a (1-1/e) ≈ 63% approximation to the adaptive optimal solution. Random sets distributed independently.

16 Stochastic Min Cost Cover Adaptively get a threshold amount of value. Minimize expected number of actions. If objective is adapt-submod and monotone, we get a logarithmic approximation. [Goemans & Vondrak, LATIN ‘06] [Liu et al., SIGMOD ‘08] [Feige, JACM ‘98] [Guillory & Bilmes, ICML ‘10] c.f., Interactive Submodular Set Cover

17 Optimal Decision Trees x1x1 x2x2 x3x = = = Garey & Graham, 1974; Loveland, 1985; Arkin et al., 1993; Kosaraju et al., 1999; Dasgupta, 2004; Guillory & Bilmes, 2009; Nowak, 2009; Gupta et al., 2010 “Diagnose the patient as cheaply as possible (w.r.t. expected cost)” 1 1 0

18 Objective = probability mass of hypotheses you have ruled out. It’s Adaptive Submodular. Outcome = 1 Outcome = 0 Test x Test w Test v

19 Conclusions New structural property useful for design & analysis of adaptive algorithms Powerful enough to recover and generalize many known results in a unified manner. (We can also handle costs) Tight analyses and optimal approximation factors in many cases x1x1 x2x2 x3x