Multilevel Coordination Mechanisms for Real-Time Autonomous Agents Edmund H. Durfee (PI) Brad Clement and Pradeep Pappachan (RAs) University of Michigan.

Slides:



Advertisements
Similar presentations
Heuristic Search techniques
Advertisements

Adopt Algorithm for Distributed Constraint Optimization
Concurrency: Deadlock and Starvation Chapter 6. Deadlock Permanent blocking of a set of processes that either compete for system resources or communicate.
Multi‑Criteria Decision Making
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
an incremental version of A*
Classical Planning via Plan-space search COMP3431 Malcolm Ryan.
Imbalanced data David Kauchak CS 451 – Fall 2013.
Planning with Resources at Multiple Levels of Abstraction Brad Clement, Tony Barrett, Gregg Rabideau Artificial Intelligence Group Jet Propulsion Laboratory.
PLANNING IN AI. Determine the set of steps that are necessary to achieve a goal Some steps might be conditional, i.e., they are only taken when a set.
Best-First Search: Agendas
Planning under Uncertainty
CPSC 335 Computer Science University of Calgary Canada.
Analyzing the tradeoffs between breakup and cloning in the context of organizational self-design By Sachin Kamboj.
CoABS Grid Component: MultiLevel Coordination Agent Edmund H. Durfee (PI) Brad Clement, Pradeep Pappachan, and Jeff Cox (GSRAs) University of Michigan.
Performance of Coordinating Concurrent Hierarchical Planning Agents Using Summary Information Brad Clement and Ed Durfee University of Michigan Artificial.
A General approach to MPLS Path Protection using Segments Ashish Gupta Ashish Gupta.
Lecture 11. Matching A set of edges which do not share a vertex is a matching. Application: Wireless Networks may consist of nodes with single radios,
Hashing General idea: Get a large array
A Multi-Agent Learning Approach to Online Distributed Resource Allocation Chongjie Zhang Victor Lesser Prashant Shenoy Computer Science Department University.
Impact of Problem Centralization on Distributed Constraint Optimization Algorithms John P. Davin and Pragnesh Jay Modi Carnegie Mellon University School.
Backtracking.
Distributed Scheduling. What is Distributed Scheduling? Scheduling: –A resource allocation problem –Often very complex set of constraints –Tied directly.
Distributed Constraint Optimization * some slides courtesy of P. Modi
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Topology aggregation and Multi-constraint QoS routing Presented by Almas Ansari.
SAMANVITHA RAMAYANAM 18 TH FEBRUARY 2010 CPE 691 LAYERED APPLICATION.
Using Abstraction in Multi-Rover Scheduling Bradley J. Clement and Anthony C. Barrett Artificial Intelligence Group Jet Propulsion Laboratory {bclement,
ANTs PI Meeting, Nov. 29, 2000W. Zhang, Washington University1 Flexible Methods for Multi-agent distributed resource Allocation by Exploiting Phase Transitions.
Analysis of Algorithms
Probabilistic Reasoning for Robust Plan Execution Steve Schaffer, Brad Clement, Steve Chien Artificial Intelligence.
Disclosure risk when responding to queries with deterministic guarantees Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University.
1 Exploring Custom Instruction Synthesis for Application-Specific Instruction Set Processors with Multiple Design Objectives Lin, Hai Fei, Yunsi ACM/IEEE.
Analysis of algorithms Analysis of algorithms is the branch of computer science that studies the performance of algorithms, especially their run time.
CP Summer School Modelling for Constraint Programming Barbara Smith 2. Implied Constraints, Optimization, Dominance Rules.
Don Perugini, Dennis Jarvis, Shane Reschke, Don Gossink Decision Automation Group Command and Control Division, DSTO Distributed Deliberative Planning.
Tetris Agent Optimization Using Harmony Search Algorithm
AI Lecture 17 Planning Noémie Elhadad (substituting for Prof. McKeown)
Chapter 2) CSP solving-An overview Overview of CSP solving techniques: problem reduction, search and solution synthesis Analyses of the characteristics.
Location Planning and Analysis Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent.
Problem Reduction So far we have considered search strategies for OR graph. In OR graph, several arcs indicate a variety of ways in which the original.
Dr. Kalpakis CMSC 421, Operating Systems Deadlocks.
Performance of Distributed Constraint Optimization Algorithms A.Gershman, T. Grinshpon, A. Meisels and R. Zivan Dept. of Computer Science Ben-Gurion University.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University.
Computer Sciences Department1.  Property 1: each node can have up to two successor nodes (children)  The predecessor node of a node is called its.
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
Adversarial Search 2 (Game Playing)
Distributed cooperation and coordination using the Max-Sum algorithm
SM Sec.1 Dated 13/11/10 STRATEGY & STRUCTURE Group 3.
Dynamic Load Balancing Tree and Structured Computations.
Algorithmic and Domain Centralization in Distributed Constraint Optimization Problems John P. Davin Carnegie Mellon University June 27, 2005 Committee:
Using Abstraction to Coordinate Multiple Robotic Spacecraft Brad Clement, Tony Barrett, Gregg Rabideau Artificial Intelligence Group Jet Propulsion Laboratory.
Wolfgang Runte Slide University of Osnabrueck, Software Engineering Research Group Wolfgang Runte Software Engineering Research Group Institute.
ITEC 202 Operating Systems
Nithin Michael, Yao Wang, G. Edward Suh and Ao Tang Cornell University
Brad Clement and Ed Durfee Artificial Intelligence Laboratory
Artificial Intelligence Problem solving by searching CSC 361
Analysis and design of algorithm
Games with Chance Other Search Algorithms
Announcements Homework 3 due today (grace period through Friday)
Brad Clement and Ed Durfee Artificial Intelligence Laboratory
CIS 488/588 Bruce R. Maxim UM-Dearborn
Searching CLRS, Sections 9.1 – 9.3.
COMP60621 Fundamentals of Parallel and Distributed Systems
Brad Clement and Ed Durfee University of Michigan
Multilevel Mission Coordination for Coalition Operations
Normal Form (Matrix) Games
Presentation transcript:

Multilevel Coordination Mechanisms for Real-Time Autonomous Agents Edmund H. Durfee (PI) Brad Clement and Pradeep Pappachan (RAs) University of Michigan Update: February 2000

The Problem Networked execution infrastructure permits concurrent, asynchronous initiation of tasks Tasks originating from different sources might impose conflicting conditions on resources, services, and states of the world Exactly which conditions matter might be decided during execution, and change dynamically Identifying which (often few) conditions require coordination (typically negotiation), when, and in what (temporal) combinations is non-trivial Negotiating over everything “just in case” can be wasteful, possibly myopic, and pose scaling difficulties

A Coalition Example Joint Mission/Exercise with objectives/responsibilities distributed among multiple commands with their own human and computational agents Operational choices within a command can unintentionally (infrequently) affect what others should or even can ultimately do (e.g., friendly fire) “Grid” services should ensure that these interactions are efficiently predicted and effectively resolved Resulting joint plan should: –Preserve room for some local run-time improvisation –Support efficient (fast, parallel) execution –Avoid unnecessarily costly actions –Require realistic runtime messaging load

Main Solution Ideas Conditions to meet (on resource assignments, environmental parameters, etc.) are typically associated with plans that pursue tasks Plans can be represented hierarchically, where abstract levels summarize the (alternative) activities they encompass Abstract plans can be used to more efficiently discover potential conflicts (or lack thereof), and to guide search for details on conflicts Choosing the right level for finding and resolving conflicts can balance coordination effort with the quality of concurrent activity and the robustness of plans.

Tradeoffs coordination levels crisper coordination lower cost more flexibility

Top-Down Search temporal constraints blocked Know as little as you can about others. Use abstract resolutions to obviate deeper ones. Reasoning at abstract levels is supported by “summary information” from the deeper “and/or” plan tree

Summary Information Approach Agents individually summarize what can or must happen in a plan subtree Compare summary information to determine no coordination needed, to find coordinating commitments, or to guide deeper search Coordinate prior to execution

Exploiting Summary Information Prune inconsistent global plans (  MightSomeWay) “Expand most threats first” (EMTF) –expand subplan involved in most threats –focuses search on driving down to source of conflict “Fewest threats first” (FTF) –search plan states with fewest threats first –or subplans involved in most threats are blocked first Branch & bound - abstract solutions help prune space where cost is higher

Experimental Domain Transports must carry evacuees to safe locations Single lane routes Routes may be destroyed 4-12 locations 2-4 transport agents

Summary Information vs. FAF FAF only returned with solutions to 6 of the 21 problems.

Summary Information vs. ExCon Points plotted where FTF-EMTF found optimal solution or both found a solution of same cost. FTF-EMTF found solutions for 16 out of 21 problems; DFS-ExCon found solutions for 8 out of 21.

Coordinating at higher levels is easier This is not clear since summary information can grow exponentially up the hierarchy in the worst case. Worst case complexity of deriving summary information is O(n 2 c 2 ) for n plans in hierarchy each with O(c) conditions. Checking the consistency of an ordering of n abstract plans each with O(c) summary conditions is O(n 2 c 2 ). Independent of the level of abstraction, checking an ordering is O(b 2d c 2 ) for O(b) subplans for each plan with O(c) conditions and hierarchy depth d. Finding a consistent synchronization of actions is equivalent to resolving threats, which is shown to be NP-complete with the number of plans. The number of abstract plans grows exponentially down the hierarchy, so difficulty grows exponentially.

Temporal Constraint Modeling Create a temporal constraint network of primitive operators (across agents). Edges labeled with temporal relation vectors from inter-operator analysis. Incrementally augment network via top- down plan elaboration and resolve potential conflicts

Runtime Coordination Cycle Receive reduction of some operator P. Add operators in the reduction to the network. To resolve a previously imposed constraint  P, Q, T  do: Fix temporal relations between Q and the operators in the reduction of P. Check consistency of temporal network. Synchronize/Select operators when necessary, request new reductions (new constraints) and block those operators. Delete  P, Q, T . Unblock P and Q if they are not involved in any other unresolved constraints. Repeat steps until there are no more unresolved constraints.

Example Plans

Stepwise Growth

Further Elaboration

Final Commitments

Coordination Example (Fine Grained Coordination) WAIT TIME SIGNAL COORDINATION TIME B A B Overlaps A (Link (3, 6) is broken)

Evaluating Multiagent Plans Intelligent Coordination Algorithms must be able to evaluate candidate multiagent plans based on well defined performance metrics in order to select plans with high “quality”. Metrics for plan evaluation should be: Computationally tractable Good predictors of plan quality Able to capture important tradeoffs that agents might wish to make during plan execution We selected the following metrics for plan evaluation: Plan Cost, Plan Reliability and Plan Reward.

A Coordination Problem (Evacuation Domain) A B

Plans for Transport Agents in the Evacuation Domain A B A12 A-SHORT-EVAC A-LONG-EVAC A26 A65 A51 A21A15 A56 A65 A51 B74 B43 B-SHORT-EVAC B-LONG-EVAC B36B68B86B63 B32 B26B68 B86 B62 B23 B34 B47 A-EVAC-6-5 B-EVAC-8

Plan Cost Plan Cost estimates the cost associated with primitive operator execution for a plan. The cost of a plan operator in a hierarchical plan can be recursively computed as follows: If the operator can be reduced uniquely to a sequence of lower level operators, its cost is the sum of the costs of those operators. If the operator can be reduced in more than one way to sequences of lower level operators, its cost is defined as the average cost of each possible reduction.

Example In the example, assume that the cost of traversing an edge is $100 and the cost of destroying an edge after traversing it is $200. Cost(A-SHORT-EVAC) = Cost(A26) + Cost(A65) + Cost(A51) = = 500 Cost(A-LONG-EVAC) = Cost(A21) + Cost(A15) + Cost(A56) + Cost(A65) + Cost(A51) = = 600 Cost(A-EVAC-6-5) = (Cost(A-SHORT-EVAC) + Cost(A- LONG-EVAC))/2 = ( )/2 = 550

Plan Reliability Plan Reliability estimates the likelihood of a plan being executed successfully in a dynamic environment. Reliability is a concern if a plan operator can be reduced in more than one way to achieve a subgoal depending on the conditions that exist at runtime and one of the reductions has to be selected a priori for coordination purposes. The Reliability Quotient (RQ) of an operator reduction is computed by the number of world states that it is equipped to handle relative to the sum total of all the states that alternative reductions can handle. The greater this ratio, the greater the reliability of the reduction, assuming that any world state is equally likely to occur. Joint plans which have operators with higher reliability quotients are valued higher w.r.t. this criterion.

Example In the example, the operator B-EVAC-8 is multiply- reducible; it can be reduced to either B-SHORT-EVAC or B-LONG-EVAC depending on the conditions at runtime. If edge (3,6) is usable, the agent will prefer the reduction B-SHORT-EVAC, otherwise it will apply the reduction B- LONG-EVAC. Since reduction B-SHORT-EVAC is applicable only when (3,6) is usable and reduction B-LONG-EVAC is preferred only if (3,6) is unusable, both reductions are mutually exclusive w.r.t. their applicability. Suppose the number of states in which in which either reduction is applicable is n. With no prior knowledge of the probability that (3,6) will be usable in the future, the number of states in which either reduction will be applicable is assumed to be n/2. Hence RQ(B-SHORT-EVAC) = RQ(B-LONG-EVAC) = (n/2)/n = 0.5

Plan Reward Agents get rewards for successfully completing their plans. Rewards are assumed to be a function of plan execution time. Smaller times correspond to greater rewards and greater times to smaller rewards. Plan reward estimates the reward that a joint plan is likely to yield. Plan reward is computed by estimating the completion times for various agent plans under the constraints of a joint plan and calculating the rewards with the help of a pre-defined reward function. Multiagent plans with higher levels of concurrency are rewarded more than those with less concurrency.

Example Two different orderings of operators (from different candidate joint plans) is shown. Assume that it takes 200 min to traverse an edge. The average completion times for A26 and B-EVAC-8 are 600 and 1000 resp. in (a) and 1000 and 1200 resp. in (b). Reward function is defined as the negative of the maximum of the completion times of the individual plans. Reward for joint plan in (a) = -max(600,1000) = Reward for joint plan in (b) = -max(1000, 1200) = Therefore, (a) is preferred to (b) under this criterion. A26 B-EVAC-8 During(A26, B-EVAC-8) (a) B-EVAC-8A26 After(A26, B-EVAC-8) (b)

Selecting Multiagent Plans based on Tradeoffs Agents might wish to make tradeoffs while selecting a joint plan for execution. Examples of possible tradeoffs are: greater reward for lesser reliability, greater reliability for greater plan cost etc. The multiagent plan evaluation algorithm must be able to evaluate candidate plans based on the tradeoffs that agents wish to make. We adopt a technique used to solve multi-criteria decision problems to evaluate candidate multiagent plans based on agent tradeoffs.

Using Multi-criteria Decision Problem Solving Techniques for Evaluating Candidate Plans In MCDPs it is necessary to evaluate available alternatives w.r.t. several criteria. To solve such problems it is necessary to compute the value of the alternatives w.r.t. each criteria and also ascertain the relative importance of the criteria (tradeoffs). The Ratio-scale method that we use, takes as input a matrix (ratio-scale matrix) that captures the relative importance of the criteria. This matrix is used to derive weights corresponding to each criterion. The value of an alternative (candidate joint plan) is the sum of its values w.r.t. each criterion weighted by the weight derived for that criterion from the ratio-scale matrix.

Plan Evaluation Algorithm For each candidate joint plan, compute the Cost, Reliability and Reward measures. From these measures compute the preference score for each candidate w.r.t. each criterion. The preference score of a candidate w.r.t. a criterion is the number of candidates with lower measures w.r.t. that criterion. Using the weights (for the criteria) derived from the Ratio- scale matrix, compute the weighted sum of the preference scores for each candidate. Select the candidate joint plan with the highest weighted sum, breaking ties arbitrarily.

Experiments

Experiments (contd.) Evacuation domain with three agents. Agents A and B have to evacuate from node 7 and C has to evacuate from 9. All routes are non-shareable and each agent has several routes. Edge (4,9) is the only edge that can fail with probability p. The cost and time for traversing edges is the same except for edge (6,7) which has a higher cost. Number of different solutions possible with different tradeoffs between plan cost, reward and reliability. The graph compares three strategies S1, S2 and S3 for different values of p. The y axis represents normalized net reward (total reward - total cost for all the agents combined). The values have been normalized w.r.t. the value computed by the oracle which is assigned a uniform maximum reward of 1. S1 corresponds to Reliability > Reward > Cost, S2 corresponds to Reward > Reliability > Cost, and S3 is a random strategy.

Experimental Results At low values of p, S2 dominates S1 because the reliability measure which is heavily weighted in S1 conservatively assumes that the likelihood of edge (4,9) failing is 0.5. At higher values of p, S1 dominates the other strategies and it finds the most reliable solution and it performance matches the oracle when p = 1. S2 finds the most efficient (concurrent) solution and its performance matches the oracle when p = 0. S1 consistently outperforms the random strategy S3 except at very low values of p.

Limitations of the Algorithm The Reward measure adopts a greedy approach while computing the extent of concurrency between various plans in a joint plan by looking only at a subset of plans involved in constraints being resolved at the current stage of the hierarchical coordination process. When there are several plans in a joint plan space, the temporal orderings of a subset of plans might very well affect the completion times of plans outside that subset. The algorithm is sensitive to the preference intensities assigned to the various criteria. E.g., observed that if the preference for reliability over reward in S1 was below a threshold, the reward and cost criteria combined dominated the reliability criterion, yielding less reliable solutions. The order in which the coordination process resolves various conflicts and the individual preferences of the agents for certain plans also impacts the quality of the solutions.

Factory Domain Agents and their Tasks There are three agents: Production Manager: Processes raw parts to produce new products. Inventory Manager: Makes parts available for processing and stows finished products. Facilities Manager: Services machines. Constraints: Raw parts and finished products occupy pre-designated slots. Only one part or finished product can occupy a slot at a time. Therefore parts or finished products must be stowed in the warehouse to make room for new ones. Some machines must be freed for servicing some time during the day.

Factory Domain WAREHOUSETOOL AREA M1 M2 M3 A/AB/E B C/CD/FD PRODUCTION MANAGER FACILITIES MANAGERINVENTORY MANAGER PARTS ON FACTORY FLOOR T1 T2 T3 FINISHED PRODUCTS RAW PARTS EF ABCD AB CD

Plan for the Production Manager Production_Plan Make_ABMake_CD Process(M1,A,Ø,A’)Process(M2,A’,B,AB)Process(M2,C,Ø,C’) Process(M3,C’,D,CD) Process(M1,C’,D,CD) Process(M,P1,P2,OP) Pre: free(M) available(P1) available(P2) In : ~free(M) ~available(P1) ~available(P2) Out: free(M) available(OP) ~available(P1) ~available(P2)

Plan for the Inventory Manager Inventory_Plan Open(E)Open(F) Swap(P1,P2)  Stow(P1) & Pickup(P2) Pickup(E) Swap(E,A)Swap(E,AB)Pickup(F)Swap(F,C)Swap(F,CD) Pickup(P) Pre: ondock(P) In: ~ondock(P) Post: available(P) Stow(P) Pre: available(P) In: ~available(P) available(P) Post: ~available(P)

Plan for the Facilities Manager Service_Plan Service_M1Service_M3 Equip(M, T) Pre: ~holding(T) free(M) In : holding(T) free(M) ~free(M) Out: holding(T) free(M) Service_M2 Equip(M1,T1)Maintain(M1)Equip(M2,T2)Maintain(M2)Equip(M3,T3)Maintain(M3) Maintain(M, T) Pre: holding(M,T) free(M) In : ~free(M) free(M) holding(M,T) Out: free(M) holding(M,T)

Trading Computation for Quality M1 - A’M2 - ABM2 - C’M3 - CD Swap E, ABSwap F, CD Service M1Service M2Service M3 Time = cpu sec. M1 - A’M2 - ABM2 - C’M3 - CD Swap E, ABSwap F, CD Service M1Service M2Service M3 Time = cpu sec. M1 - A’M2 - ABM2 - C’M3 - CD Swap E, ABSwap F, CD Service M1Service M2Service M3 Time = cpu sec.

Making Things More Concrete… Need experimental domain(s) and system(s) to map qualitative characteristics into parameters and metrics Realistic domain –Easier to justify; must address aspects like-it-or-not –Generalization can be challenging; limited experimental range; knowledge-engineering effort; harder to explain Abstract domain –Lower entry cost; scale-up easier; versatility; sharability, explainability –Harder to motivate; “doomed to succeed”

Current Status Algorithms have been implemented as Grid Ready Components Experimentation on NEO and Factory domains to explore versatility/effectiveness Analytical and experimental testing of cost and effectiveness of the algorithms, especially relative to other coordination/planning search techniques Emerging techniques for evaluating alternative coordinated plans Transitioning evaluation criteria into search heuristics

Abstract NEO Testbed Java Implementation DM CORBA Computer/communication network Control/Brokering Techniques

Concurrent Hierarchical Planner HTN planning with concurrency and ability to reason at abstract levels Soundness and completeness based on formalization of summary information Exploiting summary information in search ?Experimentally compare to others (Tsuneto et. al. ‘97) ?Characterize plans where these heuristics better/worse

pre: at(A,1,3) in: at(A,1,3),  at(B,1,3),  at(B,0,3),  at(B,1,4), at(A,0,3), at(A,1,4),  at(B,0,4),  at(A,1,3) post:  at(A,1,3),  at(B,1,3),  at(B,0,3), at(A,0,4),  at(A,0,3),  at(A,1,4),  at(B,0,3),  at(B,1,4),  at(B,0,4) Summary Information must, may always, sometimes first, last external preconditions external postconditions 1,3->0,3 0,3->0,4 1,3->0,4HI pre: at(A,1,3) in: at(A,1,3),  at(B,1,3),  at(B,0,3) post: at(A,0,3),  at(A,1,3),  at(B,1,3),  at(B,0,3) A B DADA DBDB ,3->0,4HI 1,3->0,4 1,3->0,4LO pre: at(A,1,3) in: at(A,1,3),  at(B,1,3),  at(B,0,3), at(A,0,3),  at(B,0,4),  at(A,1,3) post:  at(A,1,3),  at(B,1,3),  at(B,0,3), at(A,0,4),  at(A,0,3),  at(B,0,3),  at(B,0,4) pre: at(A,0,3) in: at(A,0,3),  at(B,0,3),  at(B,0,4) post: at(A,0,4),  at(A,0,3),  at(B,0,3),  at(B,0,4)

Determining Temporal Relations CanAnyWay(relation, p sum, q sum ) - relation can hold for any way p and q can be executed MightSomeWay(relation, p sum, q sum ) - relation might hold for some way p and q can be executed CAW used to identify solutions  MSW used to identify failure CAW and  MSW improve search  CAW and MSW  must look deeper MSW identifies threats to resolve A B DADA DBDB CanAnyWay(before, p sum, q sum )  CanAnyWay(overlaps, p sum, q sum ) MightSomeWay(overlaps, p sum, q sum ) B - before O - overlaps

Example

Evaluating Multiagent Plans Intelligent Coordination Algorithms must be able to evaluate candidate multiagent plans based on well defined performance metrics in order to select plans with high “quality”. Metrics for plan evaluation should be: Computationally tractable Good predictors of plan quality Able to capture important tradeoffs that agents might wish to make during plan execution We selected the following metrics for plan evaluation: Plan Cost, Plan Reliability and Plan Reward.

Plan Cost Plan Cost estimates the cost associated with primitive operator execution for a plan. The cost of a plan operator in a hierarchical plan can be recursively computed as follows: If the operator can be reduced uniquely to a sequence of lower level operators, its cost is the sum of the costs of those operators. If the operator can be reduced in more than one way to sequences of lower level operators, its cost is defined as the average cost of each possible reduction.

Plan Reliability Plan Reliability estimates the likelihood of a plan being executed successfully in a dynamic environment. Reliability is a concern if a plan operator can be reduced in more than one way to achieve a subgoal depending on the conditions that exist at runtime and one of the reductions has to be selected a priori for coordination purposes. The reliability of an operator reduction is computed by the number of world states that it is equipped to handle relative to the sum total of all the states that alternative reductions can handle. The greater this ratio, the greater the reliability of the reduction, assuming that any world state is equally likely to occur.

Plan Reward Agents get rewards for successfully completing their plans. Rewards are assumed to be a function of plan execution time. Smaller times correspond to greater rewards and greater times to smaller rewards. Plan reward estimates the reward that a joint plan is likely to yield. Plan reward is computed by estimating the completion times for various agent plans under the constraints of a joint plan and calculating the rewards with the help of a pre-defined reward function.

Selecting Multiagent Plans based on Tradeoffs Agents might wish to make tradeoffs while selecting a joint plan for execution. Examples of possible tradeoffs are: greater reward for lesser reliability, greater reliability for greater plan cost etc. The multiagent plan evaluation algorithm must be able to evaluate candidate plans based on the tradeoffs that agents wish to make. We adopt a technique used to solve multi-criteria decision problems to evaluate candidate multiagent plans based on agent tradeoffs.

Using Multi-criteria Decision Problem Solving Techniques for Evaluating Candidate Plans In MCDPs it is necessary to evaluate available alternatives w.r.t. several criteria. To solve such problems it is necessary to compute the value of the alternatives w.r.t. each criteria and also ascertain the relative importance of the criteria (tradeoffs). The Ratio-scale method that we use, takes as input a matrix (ratio-scale matrix) that captures the relative importance of the criteria. This matrix is used to derive weights corresponding to each criterion. The value of an alternative (candidate joint plan) is the sum of its values w.r.t. each criterion weighted by the weight derived for that criterion from the ratio-scale matrix.