Monitoring rivers and lakes [IJCAI ‘07]

Slides:



Advertisements
Similar presentations
Beyond Convexity – Submodularity in Machine Learning
Advertisements

Cost-effective Outbreak Detection in Networks Jure Leskovec, Andreas Krause, Carlos Guestrin, Christos Faloutsos, Jeanne VanBriesen, Natalie Glance.
1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Joint work with Andreas Krause 1 Daniel Golovin.
Nonmyopic Active Learning of Gaussian Processes An Exploration – Exploitation Approach Andreas Krause, Carlos Guestrin Carnegie Mellon University TexPoint.
Submodularity for Distributed Sensing Problems Zeyn Saigol IR Lab, School of Computer Science University of Birmingham 6 th July 2010.
Cost-effective Outbreak Detection in Networks Jure Leskovec, Andreas Krause, Carlos Guestrin, Christos Faloutsos, Jeanne VanBriesen, Natalie Glance.
Randomized Sensing in Adversarial Environments Andreas Krause Joint work with Daniel Golovin and Alex Roper International Joint Conference on Artificial.
1 Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin and Andreas Krause.
Parallel Double Greedy Submodular Maxmization Xinghao Pan, Stefanie Jegelka, Joseph Gonzalez, Joseph Bradley, Michael I. Jordan.
Maximizing the Spread of Influence through a Social Network
Online Distributed Sensor Selection Daniel Golovin, Matthew Faulkner, Andreas Krause theory and practice collide 1.
Polynomial Time Approximation Schemes Presented By: Leonid Barenboim Roee Weisbert.
Carnegie Mellon Selecting Observations against Adversarial Objectives Andreas Krause Brendan McMahan Carlos Guestrin Anupam Gupta TexPoint fonts used in.
Near-Optimal Sensor Placements in Gaussian Processes Carlos Guestrin Andreas KrauseAjit Singh Carnegie Mellon University.
Efficient Informative Sensing using Multiple Robots
A Decentralised Coordination Algorithm for Maximising Sensor Coverage in Large Sensor Networks Ruben Stranders, Alex Rogers and Nicholas R. Jennings School.
Near-optimal Nonmyopic Value of Information in Graphical Models Andreas Krause, Carlos Guestrin Computer Science Department Carnegie Mellon University.
Sensor placement applications Monitoring of spatial phenomena Temperature Precipitation... Active learning, Experiment design Precipitation data from Pacific.
Non-myopic Informative Path Planning in Spatio-Temporal Models Alexandra Meliou Andreas Krause Carlos Guestrin Joe Hellerstein.
Near-optimal Observation Selection using Submodular Functions Andreas Krause joint work with Carlos Guestrin (CMU)
1 Budgeted Nonparametric Learning from Data Streams Ryan Gomes and Andreas Krause California Institute of Technology.
1 Efficient planning of informative paths for multiple robots Amarjeet Singh *, Andreas Krause +, Carlos Guestrin +, William J. Kaiser *, Maxim Batalin.
Nonmyopic Active Learning of Gaussian Processes An Exploration – Exploitation Approach Andreas Krause, Carlos Guestrin Carnegie Mellon University TexPoint.
Near-optimal Sensor Placements: Maximizing Information while Minimizing Communication Cost Andreas Krause, Carlos Guestrin, Anupam Gupta, Jon Kleinberg.
Efficiently handling discrete structure in machine learning Stefanie Jegelka MADALGO summer school.
(work appeared in SODA 10’) Yuk Hei Chan (Tom)
Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
Efficient Gathering of Correlated Data in Sensor Networks
Carnegie Mellon Maximizing Submodular Functions and Applications in Machine Learning Andreas Krause, Carlos Guestrin Carnegie Mellon University.
Randomized Composable Core-sets for Submodular Maximization Morteza Zadimoghaddam and Vahab Mirrokni Google Research New York.
Maximizing the Spread of Influence through a Social Network Authors: David Kempe, Jon Kleinberg, É va Tardos KDD 2003.
5 Maximizing submodular functions Minimizing convex functions: Polynomial time solvable! Minimizing submodular functions: Polynomial time solvable!
Tractable Higher Order Models in Computer Vision (Part II) Slides from Carsten Rother, Sebastian Nowozin, Pusohmeet Khli Microsoft Research Cambridge Presented.
Submodular Maximization with Cardinality Constraints Moran Feldman Based On Submodular Maximization with Cardinality Constraints. Niv Buchbinder, Moran.
A Unified Continuous Greedy Algorithm for Submodular Maximization Moran Feldman Roy SchwartzJoseph (Seffi) Naor Technion – Israel Institute of Technology.
Maximization Problems with Submodular Objective Functions Moran Feldman Publication List Improved Approximations for k-Exchange Systems. Moran Feldman,
Submodular set functions Set function z on V is called submodular if For all A,B µ V: z(A)+z(B) ¸ z(A[B)+z(AÅB) Equivalent diminishing returns characterization:
Survivable Paths in Multilayer Networks Marzieh Parandehgheibi Hyang-won Lee Eytan Modiano 46 th Annual Conference on Information Sciences and Systems.
Deterministic Algorithms for Submodular Maximization Problems Moran Feldman The Open University of Israel Joint work with Niv Buchbinder.
Cost-effective Outbreak Detection in Networks Presented by Amlan Pradhan, Yining Zhou, Yingfei Xiang, Abhinav Rungta -Group 1.
Maximizing Symmetric Submodular Functions Moran Feldman EPFL.
Distributed Optimization Yen-Ling Kuo Der-Yeuan Yu May 27, 2010.
Polyhedral Optimization Lecture 5 – Part 3 M. Pawan Kumar Slides available online
Unconstrained Submodular Maximization Moran Feldman The Open University of Israel Based On Maximizing Non-monotone Submodular Functions. Uriel Feige, Vahab.
Inferring Networks of Diffusion and Influence
Near-optimal Observation Selection using Submodular Functions
Vitaly Feldman and Jan Vondrâk IBM Research - Almaden
Moran Feldman The Open University of Israel
István Szita & András Lőrincz
Joseph E. Gonzalez Postdoc, UC Berkeley AMPLab
Optimizing Submodular Functions
Distributed Submodular Maximization in Massive Datasets
Sorting We have actually seen already two efficient ways to sort:
Example: Feature selection Given random variables Y, X1, … Xn Want to predict Y from subset XA = (Xi1,…,Xik) Want k most informative features: A*
Chapter 6. Large Scale Optimization
The Curve Merger (Dvir & Widgerson, 2008)
Coverage Approximation Algorithms
MURI Kickoff Meeting Randolph L. Moses November, 2008
Factoring distributions V X1 X2 X3 X4 X5 X6 X7 Given random variables X1,…,Xn Partition variables V into sets A and VnA as independent as possible.
Cost-effective Outbreak Detection in Networks
Submodular Maximization Through the Lens of the Multilinear Relaxation
Near-Optimal Sensor Placements in Gaussian Processes
Clustering.
Submodular Maximization in the Big Data Era
Viral Marketing over Social Networks
Sorting We have actually seen already two efficient ways to sort:
Submodular Maximization with Cardinality Constraints
Guess Free Maximization of Submodular and Linear Sums
Presentation transcript:

Monitoring rivers and lakes [IJCAI ‘07] Need to monitor large spatial phenomena Temperature, nutrient distribution, fluorescence, … NIMS Kaiser et.al. (UCLA) Color indicates actual temperature Predicted temperature Depth Location across lake Predict at unobserved locations 1 1

Monitoring water networks [J Wat Res Mgt 2008] Contamination of drinking water could affect millions of people Hach Sensor Sensors Simulator from EPA Place sensors to detect contaminations “Battle of the Water Sensor Networks” competition ~$14K Where should we place sensors to quickly detect contamination?

Example reward functions Entropy Reward[ P(X) ] = -H(X) = x P(x) log2 P(x) Expected mean squared prediction error (EMSE) Reward[ P(X) ] = -1/n s Var(Xs), Many other objectives possible and useful… 3

Properties and Examples Submodularity Properties and Examples

Key observation: Diminishing returns Placement B = {Y1,…, Y5} Placement A = {Y1, Y2} Y2 Y1 Y‘ Adding Y’ will help a lot! Adding Y’ doesn’t help much New sensor Y’ 5

Set functions Finite set V = {1,2,…,n} Function F: 2V ! R Will always assume F(;) = 0 (w.l.o.g.) Assume black-box that can evaluate F for any input A Approximate (noisy) evaluation of F is ok (e.g., [37]) 6

Submodular set functions Submodularity: B A For AµB, sB, F(A [ {s}) – F(A) ¸ F(B [ {s}) – F(B)

Set cover is submodular A={S1,S2} S1 S2 S’ F(A[{S’})-F(A) ¸ F(B[{S’})-F(B) S1 S2 S3 S’ S4 B = {S1,S2,S3,S4} 8

Example: Mutual information Given random variables X1,…,Xn F(A) = I(XA; XVnA) = H(XVnA) – H(XVnA |XA) Lemma: Mutual information F(A) is submodular 9 9

Closedness properties F1,…,Fm submodular functions on V and 1,…,m > 0 Then: F(A) = i i Fi(A) is submodular! Submodularity closed under nonnegative linear combinations! Extremely useful fact!! F(A) submodular )  P() F(A) submodular! Multicriterion optimization: F1,…,Fm submodular, i¸0 ) i i Fi(A) submodular 10

Monotonicity A set function is called monotonic if AµBµV ) F(A) · F(B) 11

Approximate maximization Greedy algorithm: Start with A0 = ; For i = 1 to k si := argmaxs F(Ai-1 [ {s}) - F(Ai-1) Ai := Ai-1 [ {si} 12 12

One reason submodularity is useful Theorem [Nemhauser et al ‘78] Greedy algorithm gives constant factor approximation F(Agreedy) ¸ (1-1/e) F(Aopt) Greedy algorithm gives near-optimal solution! For information gain: Guarantees best possible unless P = NP! [Krause & Guestrin ’05] ~63% 13 13 13 13

Temperature data from sensor network Performance of greedy Optimal Greedy Temperature data from sensor network Greedy empirically close to optimal. Why? 14

Example: Sensor Placement

Lazy greedy algorithm [Minoux ’78] First iteration as usual Keep an ordered list of marginal benefits i from previous iteration Re-evaluate i only for top element If i stays on top, use it, otherwise re-sort Benefit s(A) a a a d b b b c c e d d c e e 16 16 16

Result of lazy evaluation 1 2 3 4 5 6 7 8 9 10 100 200 300 Number of sensors selected Running time (minutes) Exhaustive search (All subsets) Naive greedy Fast greedy Lower is better 17 17

Data dependent bounds [Minoux ’78] Suppose A is candidate solution to argmax F(A) s.t. |A| · k and A* = {s1,…,sk} be an optimal solution Then F(A*) · F(A [ A*) = F(A)+i F(A[{s1,…,si})-F(A[ {s1,…,si-1}) · F(A) + i (F(A[{si})-F(A)) = F(A) + i si For each s 2 VnA, let s = F(A[{s})-F(A) Order such that 1 ¸ 2 ¸ … ¸ n Then: F(A*) · F(A) + i=1k i 18

Bounds on optimal solution [Krause et al., J Wat Res Mgt ’08] 5 10 15 20 0.2 0.4 0.6 0.8 1 1.2 1.4 Offline (Nemhauser) bound Data-dependent bound Greedy solution Water networks data Number of sensors placed Sensing quality F(A) Higher is better Submodularity gives data-dependent bounds on the performance of any algorithm 19

The pSPIEL Algorithm [K, Guestrin, Gupta, Kleinberg IPSN ‘06] pSPIEL: Efficient nonmyopic algorithm (padded Sensor Placements at Informative and cost- Effective Locations) Select starting and ending location s1 and sB Decompose sensing region into small, well-separated clusters Solve cardinality constrained problem per cluster (greedy) Combine solutions using orienteering algorithm Smooth resulting path 20 20

Locality For two sets separated by at least distance r:

Example: Sensor Placement

pSPIEL

Approximation