Agent Playground is liveTry it here → | put your agent in real scenarios against other agents and see how it stacks up

The Big Picture

Search over which robots depend on which others (an agent dependency graph) and plan multi-step moves together — this removes the single-dependency limit of prior fast methods and improves planning when robots are large or take up more space.

The Evidence

Treating planning as a search over agent dependencies (who must move for whom) lets a fast priority-based method consider multi-step conflicts with multiple robots at once. The new algorithm, called MD-PIBT, builds and searches an agent dependency graph, plans w-step candidate paths and only executes a shorter window, and requests replans from parents when deadlocks occur. With the right hyperparameter settings MD-PIBT reproduces prior priority-inheritance behavior but can significantly outperform those methods on problems with large agents; on standard scenarios it matches the best prior results. This approach echoes the Model Context Protocol (MCP) Pattern.
Not sure where to start?Get personalized recommendations
Learn More

Data Highlights

1Real-world motivation: automated warehouses can involve up to 4,000 ground robots moving in a shared space, creating heavy congestion pressure on planners.
2Planning model: prior method plans 1-step moves (w = 1); MD-PIBT plans multi-step windows of length w (with w ≥ 1) and executes the first h steps (w ≥ h ≥ 1) so it can reason ahead before committing.
3Worst-case complexity: naive search can be as large as |P|^N (all paths for N agents), but capping how many times each agent can replan to R reduces the worst-case growth to on the order of N^R, a much smaller search in practice.

What This Means

Robotics engineers and system architects running fleets (warehouses, sorting centers) — they get better handling of congestion and large-robot footprints without redesigning priority rules. Researchers and tool builders interested in multi-robot coordination can use MD-PIBT as a general framework that can reproduce prior methods or extend them to include multi-step reasoning and kinematic constraints. Mutual Verification Pattern

Key Figures

Figure 1 : Multi-Dependence PIBT (MD-PIBT) builds and searches over an Agent Dependency Graph . Left shows an scenario for planning with a window size of 3, with initial path preferences drawn. Assume all agent’s safe paths are waiting at their current location. (1) Let MD-PIBT start planning with A. A’s path conflicts with B, D, and E’s safe path, causing A to have hard dependencies on them (Def. 2 , they must find non-safe paths for A’s path to be valid). Thus, B , D , E B,D,E need to be planned next. Given multiple agents, we plan in alphabetical order. (2) When B plans, B’s path collides with C and A’s safe path. Since A is already planned, we record a soft dependency between B and A. (3-6) This logic continues until planning F. (7) F fails to find a collision-free path. When this occurs, F requires a parent (in this case C) to replan . The replan request unplans C which includes removing downstream dependencies and converting soft dependencies to C to hard dependencies. (8) Suppose that C replans by moving down, which does not intersect with F’s safe path. Then F is not included in the AgDG. (9) After planning all agents in the AgDG, we can move on to plan other agents not in the AgDG (not depicted).
Fig 1: Figure 1 : Multi-Dependence PIBT (MD-PIBT) builds and searches over an Agent Dependency Graph . Left shows an scenario for planning with a window size of 3, with initial path preferences drawn. Assume all agent’s safe paths are waiting at their current location. (1) Let MD-PIBT start planning with A. A’s path conflicts with B, D, and E’s safe path, causing A to have hard dependencies on them (Def. 2 , they must find non-safe paths for A’s path to be valid). Thus, B , D , E B,D,E need to be planned next. Given multiple agents, we plan in alphabetical order. (2) When B plans, B’s path collides with C and A’s safe path. Since A is already planned, we record a soft dependency between B and A. (3-6) This logic continues until planning F. (7) F fails to find a collision-free path. When this occurs, F requires a parent (in this case C) to replan . The replan request unplans C which includes removing downstream dependencies and converting soft dependencies to C to hard dependencies. (8) Suppose that C replans by moving down, which does not intersect with F’s safe path. Then F is not included in the AgDG. (9) After planning all agents in the AgDG, we can move on to plan other agents not in the AgDG (not depicted).
(a) random-32-32-20
Fig 2: (a) random-32-32-20
Figure 3
Fig 3: Figure 3
(a) random-128-128-PMLA
Fig 4: (a) random-128-128-PMLA

Ready to evaluate your AI agents?

Learn how ReputAgent helps teams build trustworthy AI through systematic evaluation.

Learn More

Yes, But...

MD-PIBT introduces several hyperparameters (planning window size, execution window, collision caps, ordering policies) that need tuning per problem; defaults reproduce prior methods but best settings vary. The worst-case theoretical cost is still exponential in possible paths, so pathological layouts (long dead-end chains) can be expensive without limiting replans. Empirical gains are clearest for large-agent scenarios; on many standard small-agent benchmarks MD-PIBT matches rather than dramatically improves prior fast algorithms. The framework can be aligned with the Orchestrator-Worker Pattern.

Methodology & More

MD-PIBT reframes priority-inheritance planning as a search over agent dependencies instead of only single-step conflicts. Each agent keeps a set of candidate w-step paths and a safe fallback path (initially, waiting). The algorithm builds an agent dependency graph: if agent A’s candidate path intersects agent B’s safe path, A depends on B and B must be planned (or inherit priority) to validate A’s move. Because dependencies form a graph (not a simple chain), MD-PIBT uses a priority queue to select which dependent agent to plan next, adds hard or soft dependency edges as paths are chosen, and requests parent replans when a dead end is reached. Operationally MD-PIBT iterates: pick an agent from the dependency queue, try its next-best w-step path that doesn't collide with the current reservations (tentative paths of planned agents plus safe paths of others), add dependencies for any collisions, and either accept a path or trigger a replan cascade up the graph. The framework has hyperparameters (window sizes, collision thresholds, ordering policies) so it can reduce to standard one-step priority inheritance or extend to multi-step variants. Experiments across multiple motion models (omnidirectional, large-agent footprints, rotation-aware, and differential-drive execution) show MD-PIBT preserves the strong performance of prior methods on typical benchmarks and yields clear improvements when agents are large and multi-step interactions matter. Because MD-PIBT separates the dependency-level search from the low-level path enumeration, it is flexible to incorporate kinematic checks or serve as a collision guard for learned multi-step planners. Hierarchical Multi-Agent Pattern and A2A Protocol Pattern.
Avoid common pitfallsLearn what failures to watch for
Learn More
Credibility Assessment:

All authors affiliated with Carnegie Mellon University (a top university). h-index values are modest but strong institutional backing warrants a high credibility (4 stars).