Solving Multiagent Planning Problems with Concurrent Conditional Effects

06/19/2019 ∙ by Daniel Furelos-Blanco, et al. ∙ Universitat Pompeu Fabra Imperial College London 0

In this work we present a novel approach to solving concurrent multiagent planning problems in which several agents act in parallel. Our approach relies on a compilation from concurrent multiagent planning to classical planning, allowing us to use an off-the-shelf classical planner to solve the original multiagent problem. The solution can be directly interpreted as a concurrent plan that satisfies a given set of concurrency constraints, while avoiding the exponential blowup associated with concurrent actions. Our planner is the first to handle action effects that are conditional on what other agents are doing. Theoretically, we show that the compilation is sound and complete. Empirically, we show that our compilation can solve challenging multiagent planning problems that require concurrent actions.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.


Concurrent multiagent planning is a branch of multiagent planning in which several agents collaborate to solve a given problem. Collaboration takes the form of concurrent or joint actions that are executed together by multiple agents. Concurrent multiagent planning is challenging for several reasons: the number of concurrent actions is worst-case exponential in the number of agents, and restrictions are needed to ensure that concurrent actions are well-formed. Usually, these restrictions take the form of concurrency constraints [Boutilier and Brafman2001, Crosby2013], which model both the case for which two actions must occur in parallel, and for which they cannot occur in parallel.

In spite of recent progress in multiagent planning, there are relatively few multiagent planners that can reliably handle concurrency. CMAP [Borrajo2013], MAPlan [Stolba, Fiser, and Komenda2016] and MH-FMAP [Torreño, Onaindia, and Sapena2014] can all produce concurrent plans, but are not designed to handle more complex concurrency constraints. CrosbyJR14 (CrosbyJR14) associate concurrency constraints with the objects of a multiagent planning problem and transform the problem into a sequential, single-agent problem that can be solved using a classical planner. ShekharB18 (ShekharB18) adapt this approach using collaborative actions, i.e. single actions that involve the minimum number of agents necessary to perform a given task. BrafmanZoran14 (BrafmanZoran14) extend the distributed forward-search planner MAFS [Nissim and Brafman2014] to support concurrency constraints while preserving privacy. MaliahBS17 (MaliahBS17) propose MAFBS, which extends MAFS to use forward and backward messages.

In this paper we describe a planner that can handle arbitrary concurrency constraints, as originally proposed by BoutilierB01 (BoutilierB01) and later extended by kovacs2012multi (kovacs2012multi). Our approach is similar to previous approaches in that we transform a multiagent planning problem into a single-agent problem with much fewer actions, avoiding the exponential blowup associated with concurrent actions. The concurrency constraints of BoutilierB01 are significantly more expressive than those of Cr13b (Cr13b), enabling us to solve multiagent problems with more complex interactions (e.g. effects that depend on the concurrent actions of other agents). We show that our planner is sound and complete, and perform experiments in several concurrent multiagent planning domains to evaluate its performance.

The remainder of this paper is structured as follows. We first introduce the planning formalisms that we need to describe our planner. Next, we describe the compilation from multiagent planning to single-agent planning. We then present the results of experiments in several domains that require concurrency. Finally, we relate our planner to existing work in the literature, and conclude with a discussion.


In this section we describe the planning formalisms that we use: classical planning and concurrent multiagent planning.

Classical Planning

We consider the fragment of classical planning with conditional effects and negative conditions and goals. Given a fluent set , a literal is a valuation of a fluent in , where denotes that assigns true to , and that assigns false to . A literal set is well-defined if it does not assign conflicting values to any fluent , i.e. does not contain both and . Let be the set of well-defined literal sets on , i.e. the set of all partial assignments of values to fluents. Given a literal set , let be the complement of . We also define the projection of a literal set onto a subset of fluents .

A state is a well-defined literal set such that , i.e. a total assignment of values to fluents. Explicitly including negative literals in states simplifies subsequent definitions, but we often abuse notation by defining a state only in terms of the fluents that are true in , as is common in classical planning.

A classical planning problem is a tuple , where is a set of fluents, a set of actions, an initial state, and a goal condition (usually satisfied by multiple states). Each action has a precondition and a set of conditional effects . Each conditional effect has two literal sets (the condition) and (the effect).

An action is applicable in state if and only if , and the resulting (triggered) effect is given by

i.e. effects whose conditions hold in . We assume that is a well-defined literal set in for each state-action pair . The result of applying in is a new state . It is straightforward to show that if and are in , then so is .

A plan for planning problem is an action sequence that induces a state sequence such that and, for each such that , action is applicable in and generates the successor state . Plan solves if and only if , i.e. if the goal condition holds after applying in .

Concurrent Multiagent Planning

The standard definition of multiagent planning problems (MAPs) is due to BrafmanD08 (BrafmanD08). Formally, a MAP is a tuple , where is a set of agents and are disjoint sets of atomic actions of each agent. The fluent set , initial state and goal condition are defined as for classical (single-agent) planning. The definition and semantics of a plan are also identical to those for classical planning, except that is a sequence of joint actions, which we proceed to define.

Let be the set of all atomic actions. A joint action is a subset of atomic actions such that for each , i.e. each agent contributes at most one action to , implying . The precondition and effect of are defined as the union of the preconditions and effects of the constituent atomic actions:

A joint action is well-defined if and are well-defined literal sets in for each state .

In general, nothing prevents two atomic actions and of different agents from having conflicting preconditions or effects. Hence any joint action that includes both and is not well-defined. Moreover, some atomic actions may only be applicable together. For example, in the BoxPushing domain [Brafman and Zoran2014], some boxes are too heavy to push for a single agent, and a joint action is only applicable if enough agents push a box concurrently.

To ensure that joint actions are applicable and well-defined, researchers usually impose concurrency constraints on joint actions, which can be either positive or negative:

  • A positive concurrency constraint states that a subset of atomic actions must be performed concurrently.

  • A negative concurrency constraint states that a subset of atomic actions cannot be performed concurrently.

In PDDL 2.1 [Fox and Long2003], two actions and cannot be applied concurrently if has an effect on a fluent and has a precondition or effect on . This concurrency constraint requires no prior knowledge apart from the action definitions, but can only model negative concurrency.

Cr13b (Cr13b) defines concurrency constraints in the form of object affordances, i.e. integer intervals that determine how many agents can interact with an object concurrently. An object with affordance can only be manipulated by one agent, while requires manipulation by at least two and at most ten agents. This approach assumes that joint actions are well-defined whenever object affordances are satisfied. CrosbyJR14 (CrosbyJR14) extend the approach to affordances on object sets.

BoutilierB01 (BoutilierB01) proposed an alternative definition of concurrency constraints, later extended by kovacs2012multi (kovacs2012multi). The idea is to extend the preconditions of actions with other actions in addition to fluents. If an atomic action has precondition , then must be applied concurrently with , while a precondition implies that cannot be applied concurrently with . A joint action is only applicable if the concurrency constraints (i.e. preconditions) of all constituent atomic actions hold, and applicable joint actions are assumed to be well-defined.

As a side effect of the latter approach, we can also add concurrency constraints to the conditional effects of atomic actions. We illustrate this idea using the TableMover domain [Boutilier and Brafman2001], in which two agents move blocks between rooms using two alternative strategies:

  1. Pick up blocks and carry them using their arms.

  2. Put blocks on a table, carry the table together to another room, and tip the table to make the blocks fall down.

(:action lift-side
 :agent ?a - agent
 :parameters (?s - side)
   (and (at-side ?a ?s)
          (down ?s) (handempty ?a)
          (forall (?a2 - agent ?s2 - side)
            (not (lower-side ?a2 ?s2))))
   (and (not (down ?s)) (lifting ?a ?s)
          (up ?s) (not (handempty ?a ?s))
            (?b - block ?r - room ?s2 - side)
              (and (inroom Table ?r)
                     (on-table ?b) (down ?s2)
                     (forall (?a2 - agent)
                       (not (lift-side ?a2 ?s2))))
              (and (on-floor ?b) (inroom ?b ?r)
                     (not (on-table ?b)))))))
Figure 1: Definition of the TableMover action using the notation of Kovacs (concurrency constraints in bold).

Figure 1 shows the definition of the action in the notation of kovacs2012multi (kovacs2012multi), which is used by agent to lift side of the table. The precondition is that the side must be down (i.e. on the floor) and the agent cannot be holding anything. Moreover, the precondition also states that no other agent can lower side at the same time. When the action is applied, is no longer down but up, and is busy lifting . The action also has a conditional effect (represented by the clause): if some side is not lifted by any agent , then all blocks on the table fall to the floor. This conditional effect is what makes it possible to tip the table in order to implement the second strategy above.

Note that the action is defined using quantifiers. In practice, such quantifiers are compiled away, such that the resulting actions have quantifier-free preconditions and conditional effects, as in our definition of actions.

Below we extend the notation for classical planning to incorporate the concurrency constraints of BoutilierB01 (BoutilierB01). The idea is to view , the full set of atomic actions, as a set of fluents that can be true or false. We can now use a set of literals on to model a joint action : the fluents in corresponding to actions are true, while all other fluents in are false. Let denote the literal set on that encodes .

The next step is to define an extended fluent set , as well as a set of well-defined literal sets on . We can now encode a state and a joint action as an extended state , i.e. a literal set in .

To include concurrency constraints in the precondition and conditional effects of an action , we simply define the precondition and condition of each conditional effect as well-defined literal sets on extended fluents. Each effect is defined exclusively on fluents as before.

We can now define the semantics of a joint action . Concretely, satisfies the concurrency constraints if and only if the projected precondition holds in for each , . If is applicable, its precondition and effect in a state are the union of the preconditions and effects of the constituent atomic actions:

Note that the effects of atomic actions are conditional on the extended state . As before, we assume that is a well-defined literal set in for each pair of a state and an applicable joint action .

We use a small example to illustrate the notation. Consider a MAP with two agents and action sets and . The full set of actions is . Assume that and are defined as

The joint action is not applicable since the precondition of does not hold in the extended state . The joint action is applicable and results in the effect in any state . The joint action is also applicable and results in the effect in any state , since the condition in the conditional effect of holds in the extended state .

Compilations for MAPs

In this section we describe an approach to solving a MAP . The idea is to model each joint action using multiple atomic actions: one set of actions for selecting , one set of actions for applying , and one set of actions for resetting . The result is a classical planning problem such that the size of the action set is linear in , the number of atomic actions of agents.

Simulating a joint action using a sequence of atomic actions is problematic for the following reason: when applying an atomic action , we may not yet know which atomic actions will be applied by other agents. Since those other actions may be part of the precondition and conditional effects of , it becomes difficult to ensure that the concurrency constraints of are correctly enforced.

Our approach is to divide the simulation of a joint action into three phases: selection, application, and reset. In the selection phase, we use an auxiliary fluent to model that the atomic action has been selected. In the application phase, since the selection of atomic actions is known, we can substitute each action in preconditions and conditional effects with the auxiliary fluent . In the reset phase, various auxiliary fluents are reset. Note that each agent can apply at most one atomic action per time step, and agents collaborate to form joint actions whose constituent atomic actions are compatible and/or inapplicable on their own.


We describe the fluents in PDDL format, i.e. each fluent is instantiated by assigning objects to predicates.

The set of fluents includes all original fluents in , plus the following auxiliary fluents:

  • Fluents , , and modeling the phase.

  • For each agent , fluents , and that model the agent state: free to select an action, selected an action, and applied the action.

  • For each action in the action set of agent , a fluent which models that has been selected. We use to denote the subset of fluents of this type.

By simple inspection, the total number of fluents in is given by .

The initial state of the compilation is given by

i.e. the initial state on fluents in is , we are not simulating any joint action, and all agents are free to select actions. The goal condition is given by , i.e. the goal condition has to hold at the end of a joint action simulation.


For a literal set , let denote the projection of onto , followed by a substitution of the actions in with the corresponding fluents in . Note that both and are literal sets on fluents in , i.e. the dependence on actions in is removed.

The first four actions in the set allow us to switch between simulation phases, and are defined as follows:

: ,
: ,
: ,
: ,

For each action in the action set of agent , we define three new actions in : , and . These actions represent the three steps that an agent must perform during the simulation of a joint action.

The action causes to select action during the selection phase, and is defined as follows:

The precondition ensures that we are in the selection phase, that is free to select an action, and that the precondition of holds on fluents in . The effect prevents from selecting another action, and marks as selected.

The action applies the effect of in the application phase, and is defined as follows:

The precondition ensures that we are in the application phase, that was previously selected, and that all concurrency constraints in the precondition of hold. The effect is to apply all conditional effects of , where each condition is generated from by substituting each action with . Agent is also marked as done to prevent from being applied a second time.

The action resets auxiliary fluents to their original value, and is defined as follows:

The precondition ensures that we are in the reset phase and that was previously selected and applied (due to ). The effect is to make agent free to select actions again, and to mark as no longer selected.

Again, by inspection we can see that the total number of actions in is given by .


Figure 2: Initial state of a simple TableMover instance.

Figure 2 shows an example instance of TableMover in which the goal is for agents and to move block from room to room . An example of concurrent plan that solves this instance is defined as follows:

1(to-table a1 r1 s2)(pickup-floor a2 b1 r1)
2(putdown-table a2 b1 r1)
3(to-table a2 r1 s1)
4(lift-side a1 s2)(lift-side a2 s1)
5(move-table a1 r1 r2 s2)(move-table a2 r1 r2 s1)
6(lower-side a1 s2)

In this plan, agent first puts the block on the table, and then and concurrently lift each side of the table and move the table to room . Finally, lowers its side of the table, causing the table to tip and the block to fall to the floor.

The following sequence of classical actions in can be used to simulate the first joint action of the concurrent plan:

1(select-phase )
2(select-to-table a1 r1 s2)
3(select-pickup-floor a2 b1 r1)
4(apply-phase )
5(do-pickup-floor a2 b1 r1)
6(do-to-table a1 r1 s2)
7(reset-phase )
8(end-to-table a1 r1 s2)
9(end-pickup-floor a2 b1 r1)
10(finish )

We show that the compilation is both sound and complete.

Theorem 1 (Soundness).

A classical plan that solves can be transformed into a concurrent plan that solves .


When fluent is true, the only applicable action is . The only way to make true again is to cycle through the three phases and end with the action.

During the selection phase, a subset of actions are selected, causing the corresponding agents to be busy. Because of the precondition of the action, each selected action has to be applied in the application phase, and reset in the reset phase. The resulting simulated joint action is given by .

The precondition of holds since the precondition of each on fluents in is checked in the selection phase, during which no fluents in change values. The concurrency constraints of are checked in the application phase when all actions have already been selected. This also ensures that the conditional effects of are correctly applied. Finally, auxiliary fluents are cleaned in the reset phase. Hence the joint action satisfies all concurrency constraints and is correctly simulated by the corresponding action subsequence of .

Let be the concurrent plan composed of the sequence of joint actions simulated by the plan . Since solves , the goal condition holds at the end of , implying that also holds at the end of . This implies that solves . ∎

Theorem 2 (Completeness).

A concurrent plan that solves corresponds to a classical plan that solves .


Let be a joint action of the concurrent plan . We can use a sequence of actions in to simulate by selecting, applying and resetting each action among . Since is part of , its precondition and concurrency constraints have to hold, implying that the precondition and concurrency constraints of each atomic action hold. Hence the action sequence is applicable and results in the same effect as . By concatenating such action sequences for each joint action of , we obtain a plan . Since solves , the goal condition holds at the end of , implying that holds at the end of . This implies that solves . ∎


The basic compilation checks concurrency constraints in the application phase. Here we describe an extension that checks negative concurrency constraints in the selection phase, allowing a classical planner to identify inadmissible joint actions as early as possible, reducing the branching factor.

Assume that action has a negative concurrency constraint . As before, we can simulate this constraint using the fluent . However, may be selected after in the selection phase, in which case holds when selecting . To prevent inadmissible joint actions from being selected, we introduce additional fluents in the set :

  • For each action in the action set of agent , a fluent which indicates that cannot be selected.

We now redefine the action as follows:

To select , has to be false. For each negative concurrency constraint of , action adds fluent , preventing from being selected after .

With this extension, we only need to check positive concurrency constraints in the application phase. We also redefine such that fluents of type are reset in the cleanup phase, using the opposite effect of . The initial state and goal condition do not change since the new fluents are always false while no joint action is simulated.

The second extension is to impose a bound on the number of atomic actions in the selection phase, resulting in a classical planning problem . The fluent set extends with fluents , . We add counter parameters to select and reset actions so that they can respectively increment and decrement the value of the counter. Crucially, no select action is applicable when , preventing us from selecting more than actions. The benefit is to reduce the branching factor by restricting joint actions to have at most atomic actions.

We leave the following proposition without proof:

Proposition 3.

The compilation that includes both proposed extensions is sound.

Note that the compilation is not complete. For instance, consider a concurrent multiagent plan that contains a joint action involving 4 atomic actions. If , then the concurrent multiagent plan cannot be converted into an equivalent classical plan without exceeding the bound .

Experimental Results

We tested our compilations in four concurrent domains: TableMover, Maze, Workshop and BoxPushing111The code of the compilation and the domains are available at

In each domain, we used three variants of our compilations: unbounded joint action size, and joint action size bounded by and . In all variants, we used the extension that identifies negative concurrency constraints in the selection phase. The resulting classical planning problems were solved using Fast Downward [Helmert2006] in the LAMA setting [Richter and Westphal2010]. All experiments ran on Intel Xeon E5-2673 v4 @ 2.3GHz processors, with a time limit of 30 minutes and a memory limit of 8 GB.

The Maze domain [Crosby2014] consists of a grid of interconnected locations. Each agent in the maze must move from an initial location to a target location. The connection between two adjacent locations can be either a door, a bridge or a boat. A door can only be used by one agent at once, a bridge is destroyed upon first use, and a boat can only be used by two or more agents in the same direction.

The Workshop domain is a new domain in which the objective is to perform inventory in a high-security storage facility. It has the following characteristics:

  • To open a door, one agent has to press a switch while another agent simultaneously turns a key.

  • To do inventory on a pallet, one agent has to use a forklift to lift the pallet while another agent examines it (for security reasons, labels are located underneath pallets).

  • There are also actions for picking up a key, entering or exiting a forklift, moving an agent, and driving a forklift.

The BoxPushing domain [Brafman and Zoran2014] consists in a grid of interconnected cells. Agents must push boxes from one cell to another. Boxes have different sizes and require different numbers of agents to push (1, 2 or 3).

We use two algorithms for comparison: CrosbyJR14 (CrosbyJR14) and ShekharB18 (ShekharB18), which we refer to as CJR and SB respectively. Both algorithms define concurrency constraints in the form of affordances on sets of objects. For example, the affordance on the object set in Maze is defined as in CJR, representing that at least two agents have to row the boat between the same two locations at once. SB define the same affordance as , only allowing the minimum number of agents necessary to row a boat (i.e. ).

CJR do not separate atomic action selection from atomic action application. This is a problem since one of the atomic actions can delete the precondition of other atomic actions, thus canceling the formation of the joint action. For example, in the Maze domain, the action for crossing a bridge requires that the bridge exists, and destroys the bridge as an effect. Therefore, as this approach does not separate the selection from the application, this action can be done just by one agent at a time (and not by infinite agents as the problem states). The same occurs in the BoxPushing domain. Instances where a medium or a large box must be moved cannot be solved with this approach because the first agent to “push” the box will move it. Thus, the box location precondition for the other agent(s) does not hold, so the box is not moved in the end. On the other hand, SB extend CJR with mechanisms to avoid this problem, deferring effects until after all agents have applied their atomic actions.

Moreover, concurrency constraints in the form of object affordances are not as expressive as those of kovacs2012multi (kovacs2012multi):

  • Actions cannot appear in conditional effects, making it impossible to model TableMover instances (SB present results from a simplified version without blocks).

  • To define concurrency constraints, actions need at least one shared object, which is not the case in Workshop.

In experiments, we used Fast Downward in the LAMA setting to solve the instances produced by CJR and SB.

Domain Coverage Time (s.) Makespan # Grounded actions ()
2 4 CJR SB 2 4 CJR SB 2 4 CJR SB 2 4 CJR SB
Maze 20 13 8 6 11 9 361.5 444.2 145.6 195.1 216.1 47.2 22.0 11.7 77.3 67.7 41.7 69.3 27.9 156.8 108.2
10 8 6 5 7 6 250.2 575.6 170.4 228.4 323.1 48.3 25.0 12.2 79.6 69.8 39.9 67.4 26.1 119.3 102.1
10 5 2 1 4 3 539.5 - - - - 45.4 - - - - 43.9 71.8 30.0 194.3 115.1
BoxPushing 20 9 15 16 - 18 5.2 36.4 143.3 - 305.8 11.2 11.3 12.9 - 20.5 3.5 5.7 2.5 - 2.0
10 9 9 9 - 10 5.2 7.6 6.0 - 158.9 11.2 11.9 11.3 - 18.4 1.8 3.2 1.1 - 1.2
10 0 6 7 - 8 - 79.7 319.9 - 489.5 - 10.5 15 - 23.1 5.2 8.2 3.8 - 2.9
TableMover 24 15 12 15 - - 263.4 336.7 341.1 - - 58.7 59.0 61.5 - - 7.4 13.1 4.6 - -
12 10 10 11 - - 103.9 226.6 214.7 - - 63.5 62.0 64.5 - - 3.4 6.1 2.0 - -
12 5 2 4 - - 582.4 - - - - 49.0 - - - - 11.5 20.1 7.2 - -
Workshop 20 15 13 13 - - 134.3 301.4 52.5 - - 35.7 37.0 32.5 - - 18.0 31.0 11.5 - -
10 8 8 8 - - 42.8 263.3 37.1 - - 37.3 43.9 37.3 - - 7.7 13.6 4.8 - -
10 7 5 5 - - 238.8 362.3 77.1 - - 33.9 26.0 24.8 - - 28.2 48.3 18.1 - -
Table 1: Summary of results; see text for details. is the number of agents, is number of instances; time and length are averages for all planners that solved at least 5 instances. The number of grounded actions is an average over all instances.

Table 1 shows the results for the four domains. To provide an idea of how each planner behaves as a function of the number of agents, the table shows for each domain the same metrics for different numbers of agents.

In terms of coverage (i.e. number of solved instances), the compilation variant bounded to 2 performs the best (52, 61.9%). The unbounded compilation () and the variant bounded to 4 have similar coverage: 50 (59.5%) and 48 (57.1%) respectively. The performance of the variant bounded to 2 is not very good in BoxPushing for instances involving four agents because all of them require a large box to be pushed (i.e. three agents are required). Finally, SB and CJR are the approaches with the worst coverage: 27 (32.1%) and 11 (13.1%) respectively. The main reason is that they cannot solve TableMover and Workshop instances; CJR cannot solve BoxPushing instances either.

Regarding execution time, the unbounded compilation and the compilation bounded to 2 are the fastest. The higher the number of agents, the longer it takes to compute a plan.

In terms of makespan (i.e. number of joint actions), our approach obtains the shortest plans. CJR and SB obtain longer plans because they only construct joint actions associated with specific concurrency constraints. Any atomic action that can be applied on its own thus becomes a joint action of size 1. In contrast, our approach can combine atomic actions arbitrarily and compress the solution while planning.

The main reason that SB works better in BoxPushing is due to the hardcoded representation of collaborative actions that involve a minimum number of agents. For example, to push a box that requires agents to move, SB defines collaborative actions that involve exactly agents, while in our case, a joint action involving more than agents will also satisfy the concurrency constraints. This results in a larger branching factor which in turn affects the performance.

Note however that such a minimalist representation of collaborative actions is not always complete. For example, we can define a Maze instance where three agents have to use a boat to cross a stream. If we only define collaborative actions that involve the minimum number of agents needed to row a boat (i.e. 2), such an instance becomes unsolvable since no sequence of 2 agents rowing the boat in different directions is capable of moving all three agents to the other side. In contrast, our approach can generate a joint action that allows all three agents to cross the stream concurrently.

We also performed a scalability experiments in the Maze domain. We compare our approach (the variant) to the naive approach of converting a MAP into a classical problem by creating a classical action for each combination of agents. The instances consisted of (1) a 3x3 grid, (2) a set of agents with the same initial and goal locations, and (3) a single path to the goal that consists of interleaved boats and bridges.

Table 2 shows the number of grounded actions and solution time for varying numbers of agents. The naive approach cannot solve instances with 7 or more agents due to grounding, while our approach can solve instances with 100 agents.

#Agents # Grounded actions Time (s.)
Naive Naive
2 48 100 0.089 0.226
4 992 260 0.494 0.226
6 31248 484 53.864 0.354
8 - 772 - 0.535
10 - 1124 - 0.758
50 - 21604 - 41.979
100 - 83204 - 289.887
Table 2: Scalability of our approach () compared to the naive compilation in the Maze domain.

Related Work

Several other authors consider the problem of concurrent multiagent planning. BoutilierB01 (BoutilierB01) describe a partial-order planning algorithm for solving MAPs with concurrent actions, based on their formulation of concurrency constraints, but do not present any experimental results. CMAP [Borrajo2013] produces an initial sequential plan for solving a MAP, but performs a post-processing step to compress the sequential plan into a concurrent plan.

JonssonR11 (JonssonR11) present a best-response approach for MAPs with concurrent actions, where each agent attempts to improve its own part of a concurrent plan while the actions of all other agents are fixed. However, their approach only serves to improve an existing concurrent plan, and is unable to compute an initial concurrent plan. FMAP [Torreño, Onaindia, and Sapena2014] is a partial-order planner that also allows agents to execute actions in parallel, but the authors do not present experimental results for MAP domains that require concurrency.

The planner of CrosbyJR14 (CrosbyJR14) is similar to ours in that it also converts MAPs into classical planning problems. The authors only present results from the Maze domain, and concurrency constraints are defined as affordances on object sets that appear as arguments of actions. As we have seen, these concurrency constraints are not as flexible as those of BoutilierB01 (BoutilierB01).

BrafmanZoran14 (BrafmanZoran14) extended the MAFS multiagent distributed algorithm [Nissim and Brafman2014] to support actions requiring concurrency while preserving privacy. Messages are exchanged between agents in order to inform each other about the expansion of relevant states. Consequently, agents explore the search space together while preserving privacy. As pointed out by ShekharB18 (ShekharB18), it has two main problems: (1) it does not consider the issue of subsumed actions, and (2) it does not support concurrent actions that affect each others preconditions.

MaliahBS17 (MaliahBS17) proposed MAFBS, which extended MAFS to use forward and backward messages. This approach reduced the number of required messages and resulted in an increase in the privacy of agents.

ChouhanN16 (ChouhanN16, ChouhanN17) proposed a PDDL-like language for specifying problems involving required concurrency, which is very similar to the one by BoutilierB01 (BoutilierB01). Their planner does not make assumptions on the number of agents required to perform a joint action; rather, the number of agents is determined from the capability of agents and the objects they are interacting with. For example, in a robot domain, the number of robots required to lift an specific object can depend on the weight of the object.

ShekharB18 (ShekharB18) extended the planner of CrosbyJR14 (CrosbyJR14). Thus, it is also based on compiling the multiagent problem into a classical problem. With respect to previous work, they added collaborative actions and removed all collaborative actions that are subsumed by others (i.e. that do not involve a minimum number of agents). Besides, they showed that their approach can also be used in a distributed privacy preserving planner.

Compilations from multiagent to classical planning have also been considered by other authors. muise-codmap15 (muise-codmap15) proposed a transformation to respect privacy among agents. The resulting classical planning problem was then solved using a centralized classical planner as in our approach. Besides, compilations to classical planning have also been used in temporal planning, obtaining state-of-the-art results in many of the International Planning Competition domains [Jiménez, Jonsson, and Palacios2015].


This paper proposes a new compilation for concurrent multiagent planning problems. As far as we know, our algorithm is the first to handle concurrent conditional effects. In experiments we show that our approach is competitive with previous work, and that it can solve concurrent multiagent planning problems that are out of reach of previous approaches.

Since the number of atomic actions is exponentially smaller than the number of joint actions, a distributed action definition has the potential to scale to much larger instances, which we demonstrate in our experiments. It is not always easy to determine beforehand how many joint actions are needed; in Maze, we may need agents to cross a bridge together, requiring joint actions for agents.

In future work, we would like to explore strategies for optimizing the makespan, improving scalability and introducing the notion of capability [Chouhan and Niyogi2017]. We also want to automatically derive the bounds of our algorithm. Furthermore, privacy preserving is a central topic on multiagent planning; thus, this approach could be combined with suitable privacy-preserving mechanisms in the future.


This work has been supported by the Maria de Maeztu Units of Excellence Programme (MDM-2015-0502). Anders Jonsson is partially supported by the grants TIN2015-67959 and PCIN-2017-082 of the Spanish Ministry of Science.


  • [Borrajo2013] Borrajo, D. 2013. Plan Sharing for Multi-Agent Planning. In DMAP 2013 - Proceedings of the Distributed and Multi-Agent Planning Workshop at ICAPS, 57–65.
  • [Boutilier and Brafman2001] Boutilier, C., and Brafman, R. I. 2001. Partial-Order Planning with Concurrent Interacting Actions. J. Artif. Intell. Res. (JAIR) 14:105–136.
  • [Brafman and Domshlak2008] Brafman, R. I., and Domshlak, C. 2008. From One to Many: Planning for Loosely Coupled Multi-Agent Systems. In Proceedings of the Eighteenth International Conference on Automated Planning and Scheduling, ICAPS 2008, Sydney, Australia, September 14-18, 2008, 28–35.
  • [Brafman and Zoran2014] Brafman, R. I., and Zoran, U. 2014.

    Distributed Heuristic Forward Search with Interacting Actions.

    In Proceedings of the 2nd ICAPS Distributed and Multi-Agent Planning workshop (ICAPS DMAP-2014).
  • [Chouhan and Niyogi2016] Chouhan, S. S., and Niyogi, R. 2016. Multi-agent Planning with Collaborative Actions. In

    AI 2016: Advances in Artificial Intelligence - 29th Australasian Joint Conference, Hobart, TAS, Australia, December 5-8, 2016, Proceedings

    , 609–620.
  • [Chouhan and Niyogi2017] Chouhan, S. S., and Niyogi, R. 2017. MAPJA: Multi-agent planning with joint actions. Appl. Intell. 47(4):1044–1058.
  • [Crosby, Jonsson, and Rovatsos2014] Crosby, M.; Jonsson, A.; and Rovatsos, M. 2014. A Single-Agent Approach to Multiagent Planning. In ECAI 2014 - 21st European Conference on Artificial Intelligence, 18-22 August 2014, Prague, Czech Republic - Including Prestigious Applications of Intelligent Systems (PAIS 2014), 237–242.
  • [Crosby2013] Crosby, M. 2013. A Temporal Approach to Multiagent Planning with Concurrent Actions. PlanSIG.
  • [Crosby2014] Crosby, M. 2014. Multiagent Classical Planning. Ph.D. Dissertation, University of Edinburgh.
  • [Fox and Long2003] Fox, M., and Long, D. 2003. PDDL2.1: An Extension to PDDL for Expressing Temporal Planning Domains. J. Artif. Int. Res. 20(1):61–124.
  • [Helmert2006] Helmert, M. 2006. The Fast Downward Planning System. J. Artif. Intell. Res. (JAIR) 26:191–246.
  • [Jiménez, Jonsson, and Palacios2015] Jiménez, S.; Jonsson, A.; and Palacios, H. 2015. Temporal Planning With Required Concurrency Using Classical Planning. In Proceedings of the 25th International Conference on Automated Planning and Scheduling (ICAPS’15).
  • [Jonsson and Rovatsos2011] Jonsson, A., and Rovatsos, M. 2011. Scaling Up Multiagent Planning: A Best-Response Approach. In Proceedings of the 21st International Conference on Automated Planning and Scheduling, ICAPS 2011, Freiburg, Germany June 11-16, 2011.
  • [Kovacs2012] Kovacs, D. L. 2012. A Multi-Agent Extension of PDDL3.1. In Proceedings of the 3rd Workshop on the International Planning Competition (IPC), 19–27.
  • [Maliah, Brafman, and Shani2017] Maliah, S.; Brafman, R. I.; and Shani, G. 2017. Increased Privacy with Reduced Communication in Multi-Agent Planning. In Proceedings of the Twenty-Seventh International Conference on Automated Planning and Scheduling, ICAPS 2017, Pittsburgh, Pennsylvania, USA, June 18-23, 2017., 209–217.
  • [Muise, Lipovetzky, and Ramirez2015] Muise, C.; Lipovetzky, N.; and Ramirez, M. 2015. MAP-LAPKT: Omnipotent Multi-Agent Planning via Compilation to Classical Planning. In Competition of Distributed and Multiagent Planners.
  • [Nissim and Brafman2014] Nissim, R., and Brafman, R. I. 2014. Distributed Heuristic Forward Search for Multi-agent Planning. J. Artif. Intell. Res. 51:293–332.
  • [Richter and Westphal2010] Richter, S., and Westphal, M. 2010. The LAMA Planner: Guiding Cost-Based Anytime Planning with Landmarks. Journal of Artificial Intelligence Research 39:127–177.
  • [Shekhar and Brafman2018] Shekhar, S., and Brafman, R. I. 2018. Representing and Planning with Interacting Actions and Privacy. In Proceedings of the Twenty-Eighth International Conference on Automated Planning and Scheduling, ICAPS 2018, Delft, The Netherlands, June 24-29, 2018., 232–240.
  • [Stolba, Fiser, and Komenda2016] Stolba, M.; Fiser, D.; and Komenda, A. 2016. Potential Heuristics for Multi-Agent Planning. In Proceedings of the Twenty-Sixth International Conference on Automated Planning and Scheduling, ICAPS 2016, London, UK, June 12-17, 2016., 308–316.
  • [Torreño, Onaindia, and Sapena2014] Torreño, A.; Onaindia, E.; and Sapena, O. 2014. FMAP: Distributed cooperative multi-agent planning. Appl. Intell. 41(2):606–626.