Presentation is loading. Please wait.

Presentation is loading. Please wait.

Joint work with Deborah Estrin, UCLA (Concave Costs)

Similar presentations


Presentation on theme: "Joint work with Deborah Estrin, UCLA (Concave Costs)"— Presentation transcript:

1 Simultaneous Optimization Ashish Goel University of Southern California
Joint work with Deborah Estrin, UCLA (Concave Costs) Adam Meyerson, Stanford/CMU (Convex Costs, Concave Utilities) 4/25/2019

2 Algorithms/Theory of Computation at USC
Len Adleman Tim Chen Ashish Goel Ming-Deh Huang Mike Waterman Applications/Variations: Arbib, Desbrun, Schaal, Sukhatme, Tavare… 4/25/2019

3 My Long Term Research Agenda
Foundations of computer science (computation/interaction/information) Combinatorial optimization Approximation algorithms Discrete Applied Probability Find interesting connections Operations research; queueing theory; stochastic processes; functional analysis; game theory Find interesting application domains Communication Networks Self-Assembly 4/25/2019

4 Simultaneous Optimization
Approximately minimize the cost of a system simultaneously for a large family of cost functions Concave cost functions correspond to economies of scale Convex cost functions measure congestion on machines, on agents, or in networks Maximizing utility/profit without knowing the profit function Allocated bandwidths to customers in a network Concave profit functions correspond to the “law” of diminishing returns 4/25/2019

5 Example of Concave Utilities: Revenue Maximization in Networks
Let x = hx1,x2, …,xNi denote the bandwidth allocated to N users in a communication network These bandwidths must satisfy some linear capacity constraints, say Ax · C Let U(x) denote the total revenue (or the utility) that the network operator can derive from the system Goal: Maximize U(x), subject to Ax · C x can be thought of as “wealth” in any resource allocation problem 4/25/2019

6 The Utility Function U Standard assumptions:
U is concave (law of diminishing returns) U is non-decreasing (more bandwidth can’t harm) U(0) = 0 We will also assume that U is symmetric in x1,x2,…,xN The corresponding optimization problem is easy to solve using Operations Research techniques (eg. interior point methods) But what if U is not known? Simultaneous optimization: Maximize U(x) simultaneously for all Utility functions U 4/25/2019

7 Why simultaneous optimization?
Often, the utility function is poorly understood, eg. Customer satisfaction Often, we might want to promote several different objectives, eg. Fairness Let us focus on fairness. Should we Maximize the average utility? Be Max-min fair (steal from the rich)? Maximize the average income of the bottom half of society? Minimize the variance? Do all of the above? 4/25/2019

8 Fair Allocation Problem: Example
How to split a pie fairly? But what if there are more constraints? Alice and Eddy want only apple pie Frank: allergic to apples Cathy: equal portions of both pies David: twice as much apple pie as lemon What is a “Fair” allocation? How do we find one? 4/25/2019

9 Simultaneous optimization and Fairness
Consider the following three allocations of bandwidths to two users ha,bi hb,ai h(a+b)/2,(a+b)/2i If f measures fairness, then intuitively, f(a,b) = f(b,a) · f((a+b)/2,(a+b)/2) Hence, f is a symmetric concave function f(0) = 0, and f non-decreasing are also reasonable restrictions ie. f is a utility function All the fairness measures we found in literature are equivalent to maximizing a utility function Simultaneous optimization also promotes fairness 4/25/2019

10 Can we do Simultaneous Optimization?
Of course not But, perhaps we can do it “approximately”? Aha!! Theoretical Computer Science! More modest goal: Find x subject to Ax · C, such that for any utility function U, U(x) is a good approximation to the maximum achievable value of U 4/25/2019

11 Approximate Majorization
Given an allocation y of bandwidths to users, let Pi(y) denote the sum of the i smallest components of y Let Pi* denote the largest possible value of Pi(y) Definition: x is said to be a-majorized if Pi(x) ¸ Pi*/a for all 1· i· N Variant of the notion of majorization Interpretation: the K poorest individuals in the allocation x are collectively at least 1/a times as rich as the K poorest individuals in any other feasible allocation 4/25/2019

12 Why Approximate Majorization?
Theorem 1: An allocation x is a-majorized if and only if U(x) is an a-approximation to the maximum possible value of U for all utility functions U ie. a-majorization results in approximate simultaneous optimization Proof invokes a classic theorem of Hardy, Littlewood, and Polya from the 1920s 4/25/2019

13 Existence Theorem 2: For the bandwidth allocation problem in networks, there exists an O(log N)-majorized solution ie. Can simultaneously approximate all utility functions up to a factor O(log N) Results extend to arbitrary linear (even convex) programs, and not just the bandwidth allocation problem 4/25/2019

14 Tractability Theorem 3: Given arbitrary linear constraints, we can find (in polynomial time) the smallest a such that an a-majorized solution exists Can also find the corresponding a-majorized solution This completes the study of approximate simultaneous optimization for linear programs [Goel, Meyerson; Unpublished] [Bhargava, Goel, Meyerson; short abstract in Sigmetrics ’01] 4/25/2019

15 Examples of Utility Functions
Min Pi(x) Sum/Average åi f(xi) where f is a uni-variate utility function Eg. Entropy, åilog (1+xi) etc. Variance is also symmetric convex Can also approximately minimize the variance  Can simultaneously approximate capitalism, communism, and many other “ism”s. 4/25/2019

16 Open Problem Distributed Algorithms?? 4/25/2019

17 Example of Concave Costs: Data Aggregation in Sensor Networks
There is a single sink and multiple sources of information Need to construct an aggregation tree Data flows from the sources to the sink along the tree When two data streams collide, they aggregate Let f(k) denote the size of k merged streams Assume f(0) = 0, f is concave, f is non-decreasing Concavity corresponds to concavity of entropy/information Canonical Aggregation Functions The amount of aggregation might depend on nature of information f is not known in advance 4/25/2019

18 Another Example of Concave Costs: Buy-at-Bulk Network Design
Single source, several sinks (consumers) of information If a link serves k sinks, its cost is f(k) Economies of scale => f is concave Also, assume that f is increasing, and f(0) = 0 Goal: Construct the cheapest distribution tree Buy-at-Bulk Network Design Assume f is not known in advance Same problem as before Multicast communication: f(k) = 1 Unicast communication: f(k) = k 4/25/2019

19 A Simple Algorithmic Trick
Jensen’s Inequality E[f(X)] · f(E[X]) for any concave function f Hence, given a fractional/multipath solution, randomized rounding can only help 2 Prob. ½ 1 Prob. ¼ 0.5 2 sink source 0.5 Prob. ¼ 2 4/25/2019

20 Notation Given graph G=(V,E) and
Cost function c : E! <+ on edges Sink t Set S of K sources Cost of supporting j users on edge e is c(e)f(j), where f is an unknown canonical aggregation function Given an aggregation tree T, CT(f) = Cost of tree T for function f C*(f) = minT{CT(f)} RT(f) = CT(f)/ C*(f) 4/25/2019

21 Problem Definition Deterministic Algorithms: Problem D
Construct a tree T and give a bound on maxf{RT(f)} Randomized Algorithms: Two possibilities Problem R1: Bound on maxf{ET[RT(f)]} Problem R2: Bound on ET[maxf{RT(f)}] Will focus on problem R2 (R2 subsumes R1) Problem R1 does not model “simultaneous” optimization: no one tree needs to be good for all canonical functions. Problem R1 can be tackled using known techniques A solution of Problem R2 is likely to result in a solution of problem D using de-randomization techniques 4/25/2019

22 Previous Work Problem is NP-Hard even when f is known
Randomized O(log K log log K) approximation for problem R1 using Bartal’s tree embeddings [Bartal ’98; Awerbuch and Azar ’97] Improved to a constant factor [Guha, Meyerson, Munagala ‘01] O(log K) approximation when f is known, but can be different for different links [Meyerson, Munagala, Plotkin ’00] 4/25/2019

23 Background: Bartal’s Result
Randomized algorithm which takes an arbitrary metric space (V,dV) as input and constructs a tree metric (V,dT) such that dV(u,v) · dT(u,v), and E[dT(u,v)] · a dV(u,v), where a = O(log n log log n) Results in an O(log K log log K) guarantee for problem R1 No obvious way to extend to problem R2 Quite complicated 4/25/2019

24 Our Results Simple Algorithm
Gives a bound of 1 + log K for problem R2 Intreresting rules of thumb Can be de-randomized using pessimistic estimators and the O(1) approximation algorithm for known f Quite technical; details omitted 4/25/2019

25 Our Algorithm: Hierarchical Matching
Find the minimum cost matching between sources The “Matching” Step cost is measured in terms of shortest path distance For each matched pair, pick one at random and discard it The “Random Selection” Step Pretend that the demand from the discarded node is moved to the remaining node If two or more sources remain, go back to step 1 At the end, take a union of all the matchings and also connect the single remaining source to the sink 4/25/2019

26 Example Demands are all 1 Sink 4/25/2019

27 Example Demands are all 1 Matching 1 4/25/2019

28 Example Random Selection Step Demands are all 2 4/25/2019

29 Example Demands are all 2 Matching 2 4/25/2019

30 Example Random Selection Step Demands are all 4 4/25/2019

31 Example Demands are all 4 Matching 3 4/25/2019

32 Example Random Selection Step Demands are all 8 4/25/2019

33 Example: The Final Solution
1 2 1 Mi = Total cost of edges in i-th matching CT(f) = åi Mi f(2i-1) 8 4 1 1 2 4/25/2019

34 Bounding the Cost: Matching Step
Ci*(f) = Cost of optimal tree for function f in the residual problem after i iterations Claim:Matching cost in step i = Mi¢f(2i-1) · Ci*(f) Sink Optimal aggregation tree for six sources (t1, t2, t3, t4,t5,t6) t4 t3 t6 t5 t1 t2 4/25/2019

35 Bounding the Cost: Random Selection
Consider any edge e in the optimum aggregation tree for function f Let k(e) be the number of sinks which use e Focus on the “Random selection” step for one matched pair (u,v) k’(e) = total demand routed on edge e after this step For each of u and v, the demand is doubled with probability ½ and becomes 0 otherwise => E[k’(e)] = k(e) By Jensen’s inequality: E[f(k’(e)] · f[E[k’(e)] = f(k(e)) 4/25/2019

36 A Bound for Problem R1 Residual cost of opt. soln. is a super-martingale E[Ci*(f)] · C*(f) Expected Matching Cost in each matching step · åiE[Ci*(f)] · C*(f) In each matching step, the number of sources goes down by half =>1+log K matching steps =>Theorem 1: E[RT(f)] · 1+log K Marginal improvement of O(log log K) over Bartal’s embeddings for this problem Not very interesting 4/25/2019

37 Bound for Problem R2 Atomic aggregation functions
Ai(x) = min{x,2i} Ai is a canonical aggregation function Main Idea: Suffices to study the performance of our algorithm just for the atomic functions Details complicated; Omitted Ai(x) 2i x 4/25/2019

38 Open Algorithmic Problems
Multicommodity version (many source-sink pairs) Preliminary Progress: Can obtain O(log n log K log log n) guarantee using Bartal’s embeddings combined with our analysis Lower Bounds? Conjecture: W(log K) Handle arbitrary demands at sinks Our algorithm yields 1 + log K + log D guarantee for problem R2 where D is the maximum demand 4/25/2019

39 Open Modeling Problems
Realistic models of more general aggregation functions Information cancellation One node senses a pest infestation and sends an alarm. Another node senses high pesticide levels in the atmosphere, and sends another alarm. An intermediate node might receive both pieces of information and suppress both alarms. Amount of aggregation may depend on the set of nodes being aggregated rather than just the number Concave function f(Se) as opposed to f(ke) Bartal’s algorithm still gives an O(log K log log K) guarantee for problem R1 4/25/2019

40 Moral Why settle for one cost function when you can approximate them all? Argument against approximate modeling of aggregation functions Particularly useful for poorly understood or inherently multi-criteria problems “Information independent” aggregation 4/25/2019


Download ppt "Joint work with Deborah Estrin, UCLA (Concave Costs)"

Similar presentations


Ads by Google