
Ultimately, the DQN algorithm in Approach-2
provided the best results, showing that RL can ef-
fectively optimize collision avoidance and return ma-
noeuvres for low-thrust satellites.
ACKNOWLEDGMENT
We thank the International Astronautical Congress,
IAC 2024, Milan, Italy, October 14-18, 2024, for
feedback offered on a preliminary form of this work.
This research is partially supported by the project
“Romanian Hub for Artificial Intelligence - HRIA”,
Smart Growth, Digitization and Financial Instru-
ments Program, 2021-2027, MySMIS no. 334906
and a grant of the Ministry of Research, Innovation
and Digitization, CNCS/CCCDI-UEFISCDI, project
no. PN-IV-P8-8.1-PRE-HE-ORG-2023-0081, within
PNCDI IV.
REFERENCES
B. Gaudet, R. Linares, R. F. (2020). Six degree-of-freedom
body-fixed hovering over unmapped asteroids via li-
dar altimetry and reinforcement meta-learning. Acta
Astronautica, 172:90–99.
Boscolo Fiore, N. (2021). Machine Learning based Satellite
Collision Avoidance strategy. PhD thesis, Politecnico
Milano.
Casas, C. M., Carro, B., and Sanchez-Esguevillas, A.
(2022). Low-thrust orbital transfer using dynamics-
agnostic reinforcement learning.
D. M. Novak, M. V. (2011). Improved shaping approach to
the preliminary design of low-thrust trajectories. Jour-
nal of Guidance, Control, and Dynamics.
Gaudet, B., Linares, R., and Furfaro, R. (2020). Adaptive
guidance and integrated navigation with reinforce-
ment meta-learning. Acta Astronautica, 169:180–190.
Holt, H., Armellin, R., Baresi, N., Hashida, Y., Turconi, A.,
Scorsoglio, A., and Furfaro, R. (2021). Optimal q-
laws via reinforcement learning with guaranteed sta-
bility. Acta Astronautica, 187:511–528.
Kolosa, D. S. (2019). A Reinforcement Learning Approach
to Spacecraft Trajectory Optimization. PhD thesis,
Western Michigan University.
LaFarge, N. B., Miller, D., Howell, K. C., and Linares, R.
(2021). Autonomous closed-loop guidance using re-
inforcement learning in a low-thrust, multi-body dy-
namical environment. Acta Astronautica, 186:1–23.
Mnih, V. et al. (2013). Playing atari with deep reinforce-
ment learning. https://arxiv.org/abs/1312.5602.
N. Bourriez, A. Loizeau, A. F. A. (2023). Spacecraft au-
tonomous decision-planning for collision avoidance :
a reinforcement learning approach. 74th INTERNA-
TIONAL ASTRONAUTICAL CONGRESS (IAC).
Pinto, F. et al. (2020). Towards automated satellite con-
junction management with bayesian deep learning.
Proceedings of NeurIPS 2020, AI for Earth Sciences
Workshop.
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and
Klimov, O. (2017). Proximal policy optimization al-
gorithms. Arxiv.org/abs/1707.06347.
Sutton, R. S. and Barto, A. G. (2018). Reinforcement Learn-
ing: An Introduction. MIT Press, Cambridge, MA,
2nd edition.
Tipaldi, M., Iervolino, R., and Massenio, P. R. (2022). Rein-
forcement learning in spacecraft control applications:
Advances, prospects, and challenges. Annual Reviews
in Control, 54:1–23.
Whiffen, G. (2006). Mystic: Implementation of the
static dynamic optimal control algorithm for high-
fidelity, low-thrust trajectory design. Proceedings of
AIAA/AAS Astrodynamics Specialist Conference and
Exhibit.
Yang, C., Zhang, H., and Gao, Y. (2021). Analysis of
a neural-network-based adaptive controller for deep-
space formation flying. Advances in Space Research,
68(1):54–70.
Zavoli, A. and Federici, L. (2021). Reinforcement learn-
ing for robust trajectory design of interplanetary mis-
sions. Journal of Guidance, Control, and Dynamics,
44(8):1440–1453.
ICAART 2025 - 17th International Conference on Agents and Artificial Intelligence
1016