
the approach achieves efficient navigation and en-
hanced coverage, outperforming traditional single-
method approaches. Our results show that coverage
ranges from approximately 86.3% with one agent to
94.5% with five agents, demonstrating significant im-
provements with increasing the number of agents.
As a part of future work, we aim to extend the
proposed algorithm to handle environments where the
obstacles are moving.
ACKNOWLEDGMENT
The second author was in part supported by a research
grant from Google.
REFERENCES
Agmon, N., Hazon, N., and Kaminka, G. A. (2006). Con-
structing spanning trees for efficient multi-robot cov-
erage. In Proceedings 2006 IEEE International Con-
ference on Robotics and Automation, 2006. ICRA
2006., pages 1698–1703. IEEE.
Chibin, Z., Xingsong, W., and Yong, D. (2008). Complete
coverage path planning based on ant colony algorithm.
In 2008 15th International Conference on Mechatron-
ics and Machine Vision in Practice, pages 357–361.
IEEE.
Din, A., Ismail, M. Y., Shah, B., Babar, M., Ali, F., and
Baig, S. U. (2022). A deep reinforcement learning-
based multi-agent area coverage control for smart
agriculture. Computers and Electrical Engineering,
101:108089.
Galceran, E. and Carreras, M. (2013). A survey on coverage
path planning for robotics. Robotics and Autonomous
systems, 61(12):1258–1276.
Gronauer, S. and Diepold, K. (2022). Multi-agent deep re-
inforcement learning: a survey. Artificial Intelligence
Review, 55(2):895–943.
Guruprasad, K., Wilson, Z., and Dasgupta, P. (2012). Com-
plete coverage of an initially unknown environment
by multiple robots using voronoi partition. In Interna-
tional Conference on Advances in Control and Opti-
mization in Dynamical Systems.
Karapetyan, N., Benson, K., McKinney, C., Taslakian, P.,
and Rekleitis, I. (2017). Efficient multi-robot cover-
age of a known environment. In 2017 IEEE/RSJ In-
ternational Conference on Intelligent Robots and Sys-
tems (IROS), pages 1846–1852. IEEE.
Ladosz, P., Weng, L., Kim, M., and Oh, H. (2022). Ex-
ploration in deep reinforcement learning: A survey.
Information Fusion, 85:1–22.
Li, L., Shi, D., Jin, S., Kang, Y., Xue, C., Zhou, X.,
Liu, H., and Yu, X. (2022). Complete coverage
problem of multiple robots with different velocities.
International Journal of Advanced Robotic Systems,
19(2):17298806221091685.
Li, Y. (2017). Deep reinforcement learning: An overview.
arXiv preprint arXiv:1701.07274.
Mannadiar, R. and Rekleitis, I. (2010). Optimal coverage
of a known arbitrary environment. In 2010 IEEE In-
ternational conference on robotics and automation,
pages 5525–5530. IEEE.
Nair, V. G. and Guruprasad, K. (2020). Mr-simexcoverage:
Multi-robot simultaneous exploration and coverage.
Computers & Electrical Engineering, 85:106680.
Piardi, L., Lima, J., Pereira, A. I., and Costa, P. (2019). Cov-
erage path planning optimization based on q-learning
algorithm. In Aip conference proceedings, volume
2116. AIP Publishing.
Rekleitis, I., New, A. P., Rankin, E. S., and Choset, H.
(2008). Efficient boustrophedon multi-robot cover-
age: an algorithmic approach. Annals of Mathematics
and Artificial Intelligence, 52:109–142.
Sanghvi, N. and Niyogi, R. (2024). Distributed coverage
algorithm using multiple robots in an unknown envi-
ronment.
Sanghvi, N., Niyogi, R., and Milani, A. (2024). Sweeping-
based multi-robot exploration in an unknown environ-
ment using webots. In ICAART (1), pages 248–255.
Sehgal, A., La, H., Louis, S., and Nguyen, H. (2019). Deep
reinforcement learning using genetic algorithm for pa-
rameter optimization. In 2019 Third IEEE Interna-
tional Conference on Robotic Computing (IRC), pages
596–601. IEEE.
Senthilkumar, K. and Bharadwaj, K. K. (2012). Multi-
robot exploration and terrain coverage in an unknown
environment. Robotics and Autonomous Systems,
60(1):123–132.
Sharma, S. and Tiwari, R. (2016). A survey on multi
robots area exploration techniques and algorithms.
In 2016 International Conference on Computational
Techniques in Information and Communication Tech-
nologies (ICCTICT), pages 151–158. IEEE.
Such, F. P., Madhavan, V., Conti, E., Lehman, J., Stanley,
K. O., and Clune, J. (2017). Deep neuroevolution: Ge-
netic algorithms are a competitive alternative for train-
ing deep neural networks for reinforcement learning.
arXiv preprint arXiv:1712.06567.
Wang, Y., He, Z., Cao, D., Ma, L., Li, K., Jia, L., and Cui, Y.
(2023). Coverage path planning for kiwifruit picking
robots based on deep reinforcement learning. Com-
puters and Electronics in Agriculture, 205:107593.
DeepGen: A Deep Reinforcement Learning and Genetic Algorithm-Based Approach for Coverage in Unknown Environment
571