CS584 - Software Multiagent Systems Lecture 12 Distributed constraint optimization II: Incomplete algorithms and recent theoretical results.

Slides:



Advertisements
Similar presentations
Impact of Interference on Multi-hop Wireless Network Performance Kamal Jain, Jitu Padhye, Venkat Padmanabhan and Lili Qiu Microsoft Research Redmond.
Advertisements

Hadi Goudarzi and Massoud Pedram
Distributed Constraint Optimization Problems M OHSEN A FSHARCHI.
Adopt Algorithm for Distributed Constraint Optimization
Linear Programming Problem. Introduction Linear Programming was developed by George B Dantzing in 1947 for solving military logistic operations.
Linear Programming Problem
Towards a Theoretic Understanding of DCEE Scott Alfeld, Matthew E
1 University of Southern California Keep the Adversary Guessing: Agent Security by Policy Randomization Praveen Paruchuri University of Southern California.
Coverage by Directional Sensors Jing Ai and Alhussein A. Abouzeid Dept. of Electrical, Computer and Systems Engineering Rensselaer Polytechnic Institute.
An Approximate Truthful Mechanism for Combinatorial Auctions An Internet Mathematics paper by Aaron Archer, Christos Papadimitriou, Kunal Talwar and Éva.
Exact Inference in Bayes Nets
Optical Networks BM-UC Davis122 Part III Wide-Area (Wavelength-Routed) Optical Networks – 1.Virtual Topology Design 2.Wavelength Conversion 3.Control and.
Onur G. Guleryuz & Ulas C.Kozat DoCoMo USA Labs, San Jose, CA 95110
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
Basic Feasible Solutions: Recap MS&E 211. WILL FOLLOW A CELEBRATED INTELLECTUAL TEACHING TRADITION.
Planning under Uncertainty
DCOPs Meet the Real World: Exploring Unknown Reward Matrices with Applications to Mobile Sensor Networks Manish Jain Matthew E. Taylor Makoto Yokoo MilindTambe.
Design and Analysis of Algorithms
Dealing with NP-Complete Problems
1 University of Freiburg Computer Networks and Telematics Prof. Christian Schindelhauer Distributed Coloring in Õ(  log n) Bit Rounds COST 293 GRAAL and.
1 University of Southern California Security in Multiagent Systems by Policy Randomization Praveen Paruchuri, Milind Tambe, Fernando Ordonez University.
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Multiagent Planning with Factored MDPs Carlos Guestrin Daphne Koller Stanford University Ronald Parr Duke University.
Opportunistic Optimization for Market-Based Multirobot Control M. Bernardine Dias and Anthony Stentz Presented by: Wenjin Zhou.
Distributed Combinatorial Optimization
(work appeared in SODA 10’) Yuk Hei Chan (Tom)
Impact of Problem Centralization on Distributed Constraint Optimization Algorithms John P. Davin and Pragnesh Jay Modi Carnegie Mellon University School.
Distributed Scheduling. What is Distributed Scheduling? Scheduling: –A resource allocation problem –Often very complex set of constraints –Tied directly.
Distributed Constraint Optimization * some slides courtesy of P. Modi
Decentralised Coordination of Mobile Sensors School of Electronics and Computer Science University of Southampton Ruben Stranders,
Linear Programming David Kauchak cs161 Summer 2009.
A1A1 A4A4 A2A2 A3A3 Context-Specific Multiagent Coordination and Planning with Factored MDPs Carlos Guestrin Shobha Venkataraman Daphne Koller Stanford.
Fast Spectrum Allocation in Coordinated Dynamic Spectrum Access Based Cellular Networks Anand Prabhu Subramanian*, Himanshu Gupta*,
MAKING COMPLEX DEClSlONS
Roman Keeney AGEC  In many situations, economic equations are not linear  We are usually relying on the fact that a linear equation.
Approximation Algorithms for Stochastic Combinatorial Optimization Part I: Multistage problems Anupam Gupta Carnegie Mellon University.
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
Network Aware Resource Allocation in Distributed Clouds.
Yossi Azar Tel Aviv University Joint work with Ilan Cohen Serving in the Dark 1.
SoftCOM 2005: 13 th International Conference on Software, Telecommunications and Computer Networks September 15-17, 2005, Marina Frapa - Split, Croatia.
Distributed Constraint Optimization: Approximate Algorithms Alessandro Farinelli.
Linear Programming Data Structures and Algorithms A.G. Malamos References: Algorithms, 2006, S. Dasgupta, C. H. Papadimitriou, and U. V. Vazirani Introduction.
Heuristic Optimization Methods Greedy algorithms, Approximation algorithms, and GRASP.
Yang Cai Oct 06, An overview of today’s class Unit-Demand Pricing (cont’d) Multi-bidder Multi-item Setting Basic LP formulation.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
MDPs (cont) & Reinforcement Learning
Performance of Distributed Constraint Optimization Algorithms A.Gershman, T. Grinshpon, A. Meisels and R. Zivan Dept. of Computer Science Ben-Gurion University.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
6.853: Topics in Algorithmic Game Theory Fall 2011 Constantinos Daskalakis Lecture 22.
CSCI1600: Embedded and Real Time Software Lecture 23: Real Time Scheduling I Steven Reiss, Fall 2015.
Software Multiagent Systems: CS543 Milind Tambe University of Southern California
Efficient Resource Allocation for Wireless Multicast De-Nian Yang, Member, IEEE Ming-Syan Chen, Fellow, IEEE IEEE Transactions on Mobile Computing, April.
Distributed, Self-stabilizing Placement of Replicated Resources in Emerging Networks Bong-Jun Ko, Dan Rubenstein Presented by Jason Waddle.
1 Distributed Vertex Coloring. 2 Vertex Coloring: each vertex is assigned a color.
Distributed cooperation and coordination using the Max-Sum algorithm
R. Brafman and M. Tennenholtz Presented by Daniel Rasmussen.
Approximation Algorithms based on linear programming.
Theory of Computational Complexity M1 Takao Inoshita Iwama & Ito Lab Graduate School of Informatics, Kyoto University.
Keep the Adversary Guessing: Agent Security by Policy Randomization
Impact of Interference on Multi-hop Wireless Network Performance
Heuristic Optimization Methods
Networked Distributed POMDPs: DCOP-Inspired Distributed POMDPs
Computer Science cpsc322, Lecture 14
Multi-Agent Exploration
The story of distributed constraint optimization in LA: Relaxed
CASE − Cognitive Agents for Social Environments
Integer Programming (정수계획법)
Distributed Algorithms for DCOP: A Graphical-Game-Based Approach
Normal Form (Matrix) Games
Presentation transcript:

CS584 - Software Multiagent Systems Lecture 12 Distributed constraint optimization II: Incomplete algorithms and recent theoretical results

University of Southern California 2 Personal Assistant Agents for Scheduling Sensor Networks Multi- Spacecraft Coordination Distributed Constraint Optimization DCOP (Distributed Constraint Optimization Problem) –Agents cooperate to assign values to variables, subject to constraints, to maximize a global objective. –R(a)=∑ R S (a) for all constraints S

University of Southern California 3 Algorithms for DCOP Complete algorithms –Pro: finds optimal solution –Con: not always feasible (exponential in time or space) Local algorithms (today) –Pro: usually finds a high-quality solution quickly –Con: not optimal (but can guarantee within % of optimal)

Part I Algorithms

University of Southern California 5 k-optimality k-optimal solution (k-optimum) –No deviation by ≤ k agents can increase solution quality –Local optimum –Globally optimal solution: 000. –1-optimal solutions: 000, 111 –k-optimal algorithms: DSA (k=1), MGM-2 (k=2),...

University of Southern California 6 Approach Decompose DCOP. Each agent only sees its local constraints. Agents maximize individual utility. –(individual view of team utility)

University of Southern California 7 1-optimal algorithms Algorithms for 1 agent: –Broadcast current value to all neighbors and receive neighbors’ values. –Find new value that gives highest gain in utility, assuming that neighbors stay fixed. –Decide whether or not to change value, and act accordingly.

Algorithms 2 Variations –MGM (Maximum Gain Message) Modified DBA (Yokoo) Exchange messages with neighbors about utility gains. Neighbor that can make highest gain gets to move. –DSA (Distributed Stochastic Algorithm) Fitzpatrick and Meertens Move according to probability p. University of Southern California 8

9 x1x2x x1x2x3 000 MGM – Maximal Gain Message Monotonic algo., but gets stuck in a local optimum! Only one agent in a neighborhood moves at a time

University of Southern California 10 x1x2x x1x2x DSA – Distributed Stochastic Algorithm One possible path (say p=0.5): x1x2x

University of Southern California 11 Experimental Domains Regular graph coloring (~sensor network) –Cost if neighbors choose same value. Randomized DCOP –Each combination of neighbors’ values gets uniform random reward High-stakes scenario (~UAVs) –Large cost if neighbors choose same value. –Otherwise, small uniform random reward is given. –Add “safe” value where all agents start. No reward or penalty if neighbors choose this value.

University of Southern California 12 DSA vs. MGM Graph coloring and Randomized DCOP: –DSA gives higher solution quality than MGM. –DSA improves more quickly than MGM. High-stakes scenario: –DSA and MGM give same solution quality. –MGM generally improves more quickly than DSA. But, these graphs are averages....

University of Southern California 13 DSA vs. MGM MGM increases monotonically Much better for –anytime algorithm –high-stakes domains.

University of Southern California 14 Algorithms with higher k Until now (DSA, MGM), agents have only acted based on their own, local constraints –a myopic worldview Now, we look at algorithms where agents form groups, and act based on all constraints in the group. –enlarging the worldview First step: groups of 2. –“2-optimality” –Maheswaran, Pearce, and Tambe ‘04

University of Southern California 15 Coordinated Algorithms All agents are either offerers or receivers with probability q. Offerers: –Pick neighbor j at random, and calculate my gains from all combinations of values from myself and j. –Send this information (several offers) as a message to j.,…> Receivers: –Accept the offer that makes my group’s gain the highest, or just move alone instead. –groupGain = offerersGain + receiversGain - gain in common link. –If I accept an offer, tell the offerer which one I am accepting, and how much our group will gain.

University of Southern California 16 2-optimal algorithms To improve solution quality, agents can form groups of 2 Groups move according to group utility –sum of all constraints on any group member 2-optimal algorithm –any connected group of 2 agents can coordinate to make a joint move. 2-optimum –state at which no group of up to 2 agents can make a joint move that will increase group reward. Any 2-optimum is also a 1-optimum

University of Southern California 17 MGM-2 Form groups of 2 agents and then do: –Send my gain (can be group gain) to all my neighbors. –Receive gain messages from my neighbors. –If I am involved in an accepted offer, If my gain > neighbors’ gain (not counting my partner), send “yes” to my partner. If not, then send “no” to my partner. If I sent “yes” and got “yes”, then make the move in the offer. –If I am not involved in an offer If my gain > neighbors’ gain, then make my best move. 5 message cycles per move –(offer, accept, gain, confirm, move). Monotonically increasing solution quality

University of Southern California 18 x1x2x x1x2x3 MGM-2 Example x1x2x3 no gains offerer receiver offerer x1, x2 gain=7 x2, x3 gain=7 x1x2x3 accepts x1, x2 group gain=2 x1x2x3 receiver offerer receiver x2, x3 gain=12 accepts x2, x3 group gain=12

University of Southern California 19 SCA-2 (Stochastic Coordination Algorithm) Based on DSA If offerer –Send out offers to a randomly chosen neighbor. –If offer accepted, prepare to do the move in the offer. –If offer not accepted, prepare to move alone (pick move with highest individual gain). If receiver –If accepting offer, send acceptance message back to offerer, and prepare to do the move in the offer. –Else, prepare to move alone. Move, according to probability p. 3 message cycles per move (offer, accept, move).

University of Southern California 20 Experimental Trends Monotonic (1-opt, 2-opt) Stochastic (1-opt, 2-opt)

Part II Theoretical Results

University of Southern California 22 Guarantees on Solution Quality Guarantee of k-optimum as % of global optimum –Factors: k (how local of an optimum) m (maximum -arity of constraints) n (number of agents) constraint graph structure (if known) Note: actual costs/rewards on constraints –distributed among agents, not known a priori

University of Southern California 23 Guarantees on Solution Quality Three results Guarantees for: –Fully-connected DCOP graphs Applies to all graphs (i.e. when graph is unknown) Closed-form equation –Particular graphs Stars Rings Closed-form equation –Arbitrary DCOP graphs Linear program

University of Southern California 24 Fully-Connected Graph Reward of k-optimum in terms of global optimum Independent of rewards Independent of domain size Provably tight (in paper) One assumption: rewards are non-negative For binary graph (m=2),

University of Southern California 25 Proof sketch / example 10R(a) ≥ ∑R(â) 10R(a) ≥ 3R(a*) + 1R(a) Fully connected graph n = 5 agents m = 2 (binary constraints) k = 3 Goal: express R(a) in terms of R(a*). a dominates: Â = { } a* = (global opt) a = (3-opt)

University of Southern California 26 Other graph types Ring: Star: Similar analysis, but exploit graph structure –Only consider  where connected subsets of k agents deviate

University of Southern California 27 Proof sketch / example 5R(a) ≥ ∑R(â) 5R(a) ≥ 2R(a*) + 1R(a) Ring graph n = 5 agents m = 2 (binary constraints) k = 3 Goal: express R(a) in terms of R(a*). a dominates: Â = { } a* = (global opt) a = (3-opt)

University of Southern California 28 Arbitrary graph Arbitrary graph = linear program Minimize R(a)/R(a*) such that: –for all dominated assignments â, R(a) - R(â) ≥ 0. Each constraint S in DCOP = 2 variables in LP. –1: R S (a) for reward on S in k-optimal solution –2: R S (a*) for reward on S in global optimum –All other rewards on S taken as 0 (as before). Why ok? R(a) and R(a*) don’t change a still k-optimal (no k agents would change) a* still globally optimal

University of Southern California 29 Arbitrary graph Minimize R(a)/R(a*) such that: –for all dominated assignments â, R(a) - R(â) ≥ 0. R(a) = ∑ R S (a) over all constraints S. R(a*) = ∑ R S (a*) over all S. Every R(â) = ∑ R S (a) + ∑ R S (a*) for various S. Why? –Consider assignment in 5-agent ring. R(00011) = R {1,2} (00011) + R {2,3} (00011) + R {3,4} (00011) + R {4,5} (00011) + R {5,1} (00011) R {1,2} (a) + R {2,3} (a) R {4,5} (a*) + 0 Minimize ∑ R S (a) / ∑ R S (a*) –Subject to linear inequalities on R S (a) and R S (a*) for various S. Linear Fractional Program: reducible to LP (textbook)

University of Southern California 30 Experimental Results Designer can choose appropriate k or topology!

University of Southern California 31 Experimental Results

University of Southern California 32 Conclusions Guarantees for k-optima in DCOPs as % of optimum –Despite not knowing constraint rewards –Helps choose algorithm to use –Helps choose topology to use Big idea: –Single agent: Rationality -> Bounded Rationality –Multi agent: Global Optimality -> k-Optimality –Ability to centralize information (coordinate) is bounded (only groups of k agents) –Guarantees on performance of “bounded coordination”

University of Southern California 33 Readings J. P. Pearce and M. Tambe, "Quality Guarantees on k- Optimal Solutions for Distributed Constraint Optimization Problems," in IJCAI-07. R. T. Maheswaran, J. P. Pearce and M. Tambe, "Distributed Algorithms for DCOP: A Graphical-Game-Based Approach," in PDCS-04. –(just read algorithms - don’t need to read proofs) W. Zhang, Z. Xing, G. Wang and L. Wittenburg, "An analysis and application of distributed constraint satisfaction and optimization algorithms in sensor networks," in AAMAS- 03.