dicates that it has overfitted to the simulated environ-
ment.
6 DISCUSSION AND
CONCLUSION
The results show that an RL based agent is capable of
planning a headland turn and controlling the vehicle
to follow it. The final policy can easily be executed
in under 100ms, meaning that the approach is viable
for running in real time. However, the stability of the
training is still somewhat lacking. While we are able
to find a great policy even with the instability, improv-
ing on this would be desirable to speed up training and
allow for more difficult problems.
Since this paper is intended as a proof-of-concept,
no in-depth comparison to other headland turning al-
gorithms is made. This should be an important com-
ponent for future research into the area.
It is also clear that this implementation does not
use many of the theoretical advantages of RL. Re-
search that focuses on using more complex vehicle
models and optimization goals could greatly benefit
from those strengths.
For the real world application, more techniques to
improve the sim-to-real transfer could be applied. The
results could likely be improved significantly, if the
policy was forced to achieve better generalization.
ACKNOWLEDGMENTS
We would like to thank Michael Maier and Ruben
Hefele for their help with implementing and testing
the policy on the real tractor. We are also very great-
ful for Henri Hornburg’s input on possible algorithms
for the expert knowledge generation.
REFERENCES
Cariou, C., Lenain, R., Thuilot, B., and Martinet, P.
(10.10.2009 - 15.10.2009). Motion planner and
lateral-longitudinal controllers for autonomous ma-
neuvers of a farm vehicle in headland. In 2009
IEEE/RSJ International Conference on Intelligent
Robots and Systems, pages 5782–5787. IEEE.
Dubins, L. E. (1957). On Curves of Minimal Length with a
Constraint on Average Curvature, and with Prescribed
Initial and Terminal Positions and Tangents. American
Journal of Mathematics, 79(3):497.
Gao, G., Guo, H., Zhang, J., Zhang, Z., Wu, T., Lu, H.,
Qiu, Z., Chen, H., and Lingxuan, Z. (2023). An ef-
ficient headland-turning navigation system for a saf-
flower picking robot. Journal of Agricultural Engi-
neering, 54(3).
Lazzaroni, L., Pighetti, A., Bellotti, F., Capello, A., Cossu,
M., and Berta, R. (2024). Automated Parking in
CARLA: A Deep Reinforcement Learning-Based Ap-
proach. In Bellotti, F., Grammatikakis, M. D., Man-
sour, A., Ruo Roch, M., Seepold, R., Solanas, A., and
Berta, R., editors, Applications in Electronics Pervad-
ing Industry, Environment and Society, volume 1110
of Lecture Notes in Electrical Engineering, pages
352–357. Springer Nature Switzerland.
Makino, T., Yokoi, H., and Kakazu, Y. (28-30 July 1999).
Development of a motion planning system for an agri-
cultural mobile robot. In SICE ’99. Proceedings of
the 38th SICE Annual Conference. International Ses-
sion Papers (IEEE Cat. No.99TH8456), pages 959–
962. Soc. Instrum. & Control Eng.
Oksanen, T. (2007). Path planning algorithms for agricul-
tural field machines, volume no. 31 0031 of Helsinki
University of Technology Automation Technology
Laboratory Series A Research reports. Helsinki Uni-
versity of Technology, Espoo.
Olcay, E., Rui, X., and Wang, R. (26.08.2023 - 30.08.2023).
Headland Turn Automation Concept for Tractor-
Trailer System with Deep Reinforcement Learning.
In 2023 IEEE 19th International Conference on Au-
tomation Science and Engineering (CASE), pages 1–
7. IEEE.
Sabelhaus, D., R
¨
oben, F., Meyer zu Helligen, L. P., and
Schulze Lammers, P. (2013). Using continuous-
curvature paths to generate feasible headland turn ma-
noeuvres. Biosystems Engineering, 116(4):399–409.
S
´
anchez, M., Morales, J., and Mart
´
ınez, J. L. (2023). Re-
inforcement and Curriculum Learning for Off-Road
Navigation of an UGV with a 3D LiDAR. Sensors
(Basel, Switzerland), 23(6).
Schulman, J., Levine, S., Moritz, P., Jordan, M. I., and
Abbeel, P. (2015). Trust Region Policy Optimization.
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and
Klimov, O. (2017). Proximal Policy Optimization Al-
gorithms.
Tu, X. (2013). Robust navigation control and headland
turning optimization of agricultural vehicles. PhD
thesis, Iowa State University.
Wiberg, V., Wallin, E., Nordfjell, T., and Servin, M. (2022).
Control of Rough Terrain Vehicles Using Deep Rein-
forcement Learning. IEEE Robotics and Automation
Letters, 7(1):390–397.
Wu, Y., Wang, L., Lu, X., Wu, Y., and Zhang, H.
(27.10.2023 - 29.10.2023). Reinforcement Learning-
based Autonomous Parking with Expert Demonstra-
tions. In 2023 7th CAA International Conference on
Vehicular Control and Intelligence (CVCI), pages 1–
6. IEEE.
Yin, X., Wang, Y., Chen, Y., Jin, C., and Du, J. (2020).
Development of autonomous navigation controller for
agricultural vehicles. International Journal of Agri-
cultural and Biological Engineering, 13(4):70–76.
Zhou, K., Jensen, A. L., Bochtis, D. D., and Sørensen, C. G.
(2015). Quantifying the benefits of alternative field-
work patterns in a potato cultivation system. Comput-
ers and Electronics in Agriculture, 119:228–240.
Reinforcement Learning for Autonomous Headland Turns
325