r
1, successfullanechanging
5, collision
0, notexecutelanechanging
(16)
As shown in Fig. 5, the numbers of collisions for
single-step DQN and general DQN both decrease
during training. After 1,700 training episodes, there is
no more collisions happening for single-step DQN.
But general DQN cannot completely converge to 0.
This shows that our algorithm can converge better. It
can teach the autonomous vehicle to learn to judge the
feasibility of lane-changing ensuring absolute safety.
Figure 5: Training Results.
6 CONCLUSIONS
In this paper, we proposed a new method to judge the
feasibility when the autonomous vehicle is going to
change the lane. The method combines the single-step
reinforcement learning and the deep reinforcement
learning. We use the single-step reinforcement
learning framework that learns by solving the
expected reward for executing different actions in the
same state. Aiming at the problem of discontinuous
states or actions in this framework, combined with the
idea of DQN, a neural network is used to approximate
the Q value function. The proposed single-step DQN
algorithm judges the feasibility of lane-changing
based on the lane-changing plan made by the high-
layer path planning module and the surrounding
vehicle state obtained by sensors. The instruction is
sent to the low-level control module, which uses the
LQR-based method to complete the lane-changing.
The final results indicate that the proposed method in
this paper can ensure that the lane-changing process
of autonomous vehicle is absolutely safe.
ACKNOWLEDGEMENTS
This work was supported by the European Union’s
Horizon 2020 research and innovation programme
under the Marie Skłodowska-Curie grant agreement
No 824019, and Beijing Municipal Science and
Technology Commission under Grant
D17110000491701.
REFERENCES
Alizadeh, A., Moghadam, M., Bicer, Y., et al. Automated
Lane Change Decision Making using Deep
Reinforcement Learning in Dynamic and Uncertain
Highway Environment. 2019 IEEE Intelligent
Transportation Systems Conference (ITSC). October,
2019. pp. 1399-1404.
Aufrère, R., Gowdy, J., Mertz, C., et al. Perception for
collision avoidance and autonomous driving.
Mechatronics, 13.10 (2003), 1149-1161.
Bojarski, M., Del Testa, D., Dworakowski, D., et al. End to
end learning for self-driving cars. (2016). arXiv
preprint arXiv:1604.07316.
Chen, J., Zhao, P., Liang, H., et al. A multiple attribute-
based decision making model for autonomous vehicle
in urban environment. 2014 IEEE Intelligent Vehicles
Symposium Proceedings. June, 2014. pp. 480-485.
Codevilla, F., Miiller, M., López, A., et al. End-to-end
driving via conditional imitation learning. 2018 IEEE
International Conference on Robotics and Automation
(ICRA). May, 2018. pp. 1-9.
Desjardins, C., & Chaib-Draa, B. Cooperative adaptive
cruise control: A reinforcement learning approach.
IEEE Transactions on intelligent transportation
systems, (2011) 12(4), 1248-1260.
Gu, S., Holly, E., Lillicrap, T., et al. Deep reinforcement
learning for robotic manipulation with asynchronous
off-policy updates. 2017 IEEE international conference
on robotics and automation (ICRA). May, 2017. pp.
3389-3396.
Hidas, P. Modelling vehicle interactions in microscopic
simulation of merging and weaving. Transportation
Research Part C: Emerging Technologies. (2005) 13(1),
37-62.
Kamaldinov, I., & Makarov, I. Deep reinforcement learning
in match-3 game. 2019 IEEE conference on games
(CoG). August, 2019. pp. 1-4.
Li, D., Zhao, D., Zhang, Q., et al. Reinforcement learning
and deep learning based lateral control for autonomous
driving. (2018). arXiv preprint arXiv:1810.12778.
Santana, E., & Hotz, G. Learning a driving simulator.
(2016). arXiv preprint arXiv:1608.01230.
Schwarting, W., Alonso-Mora, J., & Rus, D. Planning and
decision-making for autonomous vehicles. Annual
Review of Control, Robotics, and Autonomous
Systems. (2018).
Collision Numbers
per 100 Episodes