
In Sixth International Conference on Computer Vision
(IEEE Cat. No.98CH36271), pages 59–66.
Ryan, M. R. (2008). Exploiting subgraph structure in multi-
robot path planning. Journal of Artificial Intelligence
Research, 31:497–542.
Sartoretti, G., Kerr, J., Shi, Y., Wagner, G., Kumar,
T. S., Koenig, S., and Choset, H. (2019). Primal:
Pathfinding via reinforcement and imitation multi-
agent learning. IEEE Robotics and Automation Let-
ters, 4(3):2378–2385.
Semiz, F. and Polat, F. (2021). Incremental multi-agent
path finding. Future Generation Computer Systems,
116:220–233.
Sharon, G., Stern, R., Felner, A., and Sturtevant, N. R.
(2015). Conflict-based search for optimal multi-agent
pathfinding. Artificial Intelligence, 219:40–66.
Shofer, B., Shani, G., and Stern, R. (2023). Multi agent path
finding under obstacle uncertainty. In Proceedings of
the International Conference on Automated Planning
and Scheduling, volume 33, pages 402–410.
Standley, T. (2010). Finding optimal solutions to coopera-
tive pathfinding problems. In Proceedings of the AAAI
conference on artificial intelligence, volume 24, pages
173–178.
Stern, R. (2019). Multi-agent path finding–an overview. Ar-
tificial Intelligence: 5th RAAI Summer School, Dol-
goprudny, Russia, July 4–7, 2019, Tutorial Lectures,
pages 96–115.
Stern, R., Sturtevant, N., Felner, A., Koenig, S., Ma, H.,
Walker, T., Li, J., Atzmon, D., Cohen, L., Kumar,
T., Barták, R., and Boyarski, E. (2021). Multi-agent
pathfinding: Definitions, variants, and benchmarks.
Proceedings of the International Symposium on Com-
binatorial Search, 10:151–158.
Stützle, T. and Dorigo, M. (2004). Ant Colony Optimiza-
tion. The MIT Press.
Verbari, P., Bascetta, L., and Prandini, M. (2019). Multi-
agent trajectory planning: A decentralized iterative
algorithm based on single-agent dynamic rrt*. 2019
American Control Conference (ACC), pages 1977–
1982.
Wan, Q., Gu, C., Sun, S., Chen, M., Huang, H., and Jia,
X. (2018). Lifelong multi-agent path finding in a dy-
namic environment. In 2018 15th International Con-
ference on Control, Automation, Robotics and Vision
(ICARCV), pages 875–882.
Wiedemann, T., Vlaicu, C., Josifovski, J., and Viseras, A.
(2021). Robotic information gathering with reinforce-
ment learning assisted by domain knowledge: An ap-
plication to gas source localization. IEEE Access,
9:13159–13172.
Winston, W. (2022). Operations Research: Applications
and Algorithms. Cengage Learning.
APPENDIX
The components of proposed approaches (mentioned in 7.2)
are evaluated through a series of statistical analyses to en-
sure their reliability and consistency during the evaluation
process. Firstly, normality tests using the Shapiro-Wilk test
were conducted, and all evaluation metrics (average EMD,
optimality gap, runtime, and max. concession difference)
failed across all groups. Consequently, non-parametric tests
were applied for further comparison.
For success rate, the Cochran’s Q test was used and in-
dicated no significant difference (p ≈ .313 > .05). Sim-
ilarly, the average EMD was evaluated using the Fried-
man test, which showed no significant difference (p ≈
.168 > .05). Lastly, the optimality gap was analyzed with
the Kruskal-Wallis test, yielding no significant difference
(p ≈ .994 > .05). Additionally, for both runtime and max.
concession difference, the Friedman test was applied and
showed significant differences for both metrics (p ≈ .003 <
.05 for runtime, p ≈ .000 < .05 for max. concession Differ-
ence), prompting pairwise comparisons using the Wilcoxon
signed rank test (Table 2).
Table 2: Pairwise Comparisons (p-values) for Runtime and
Max. Concession Difference.
Comparison Runtime Max. Concession Difference
Random (EECBS) vs Fair Token (EECBS) .030 .002
Probability-Based (CBSH2-RTC) vs Fair Token (EECBS) .019 .000
Fair Token (EECBS) vs Random (CBSH2-RTC) .005 .012
Fair Token (EECBS) vs Probability-Based (EECBS) .008 .000
Random (EECBS) vs Probability-Based (CBSH2-RTC) .330 .179
Random (EECBS) vs Random (CBSH2-RTC) .085 .957
Random (EECBS) vs Fair Token (CBSH2-RTC) .829 .003
Probability-Based (CBSH2-RTC) vs Random (CBSH2-RTC) .543 .092
Probability-Based (CBSH2-RTC) vs Probability-Based (EECBS) .510 .519
Probability-Based (CBSH2-RTC) vs Fair Token (CBSH2-RTC) .274 .000
Fair Token (EECBS) vs Fair Token (CBSH2-RTC) .340 .900
Random (CBSH2-RTC) vs Probability-Based (EECBS) .232 .074
Random (CBSH2-RTC) vs Fair Token (CBSH2-RTC) .178 .003
Probability-Based (EECBS) vs Fair Token (CBSH2-RTC) .513 .001
Now, the revising strategies (mentioned in 7.3) are eval-
uated through a series of statistical analyses. For all evalua-
tion metrics (optimality gap, runtime, average EMD) failed
the normality test (i.e., Shapiro-Wilk), leading to the use of
non-parametric tests again, illustrated in Table 3. The Fried-
man test was applied for runtime and average EMD all show
significant differences (p ≈ .000 < .05), followed by the
Wilcoxon signed rank test for pairwise comparisons. For
the optimality gap, the Kruskal-Wallis test indicated signifi-
cant differences (p ≈ .000 < .05), prompting Conover-Iman
test for pairwise comparisons. Finally, success rate were
analyzed using Cochran’s Q test, which also showed signif-
icant differences (p ≈ .000 < .05), followed by McNemar’s
test for pairwise comparisons.
Table 3: Pairwise Comparisons (p-values) for Revising
Strategies.
Comparison Optimality Gap Average EMD Runtime Success Rate
Enhanced ACO vs Basic ACO .000 .014 .000 .023
Enhanced ACO vs Only Waiting .000 .000 .000 .000
Basic ACO vs Only Waiting .402 .000 .000 .000
Basic ACO vs No-Repairing - - .000 .000
Enhanced ACO vs No-Repairing - - .000 .000
Only Waiting vs No-Repairing - - .000 .005
A Multitier Approach for Dynamic and Partially Observable Multiagent Path-Finding
573