curve closely resembles human driving while cross-
ing intersections. The inclusion of early stopping and
avoiding high speed at intersections ensures the car is
always in a safe state. Since we are using real values
of brake and throttle as compared to previous study by
Chae (Chae et al., 2017) where the author proposes
discrete values of deceleration, we get a smooth de-
crease/increase in velocity. Also the velocity is in the
safety limits as we have imposed penalties for high
speed.
6 CONCLUSION AND FUTURE
WORK
We have demonstrated the braking and throttle control
system for 2 scenarios. One where braking is very
important, while the other scenario needs both braking
and throttle action. DDPG is used to ensure that the
values of brake and throttle are smoothly changing in-
stead of a sudden change. We have proposed measures
to avoid early stopping and high speed movement at
an intersection, making autonomous driving similar
to human driving. This work can be further extended
to other such situations, where the vehicle must apply
brake or throttle to avoid emergency situations. An
interesting study would be to see the effect of braking
and throttle actions along with steering of the vehicle
in these situations. One can create a unified model
for all these scenarios with a singular reward func-
tion. Also, the effect of change of weather conditions
can also be considered so as to make the model more
realistic.
ACKNOWLEDGEMENTS
The authors would like to thank Dr. Prashant P. Bar-
takke, Department of Electronics and Telecommuni-
cation Engineering, College of Engineering Pune, for
guiding us throughout the project.
REFERENCES
Chae, H., Kang, C. M., Kim, B., Kim, J., Chung, C. C.,
and Choi, J. W. (2017). Autonomous braking system
via deep reinforcement learning. In 2017 IEEE 20th
International Conference on Intelligent Transportation
Systems (ITSC), pages 1–6.
Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A., and
Koltun, V. (2017). CARLA: An open urban driving
simulator. In Proceedings of the 1st Annual Conference
on Robot Learning, pages 1–16.
Fu, Y., Li, C., Yu, F. R., Luan, T. H., and Zhang, Y. (2020).
A decision-making strategy for vehicle autonomous
braking in emergency via deep reinforcement learn-
ing. IEEE Transactions on Vehicular Technology,
69(6):5876–5888.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning. The MIT Press.
Kaempchen, N., Schiele, B., and Dietmayer, K. (2009).
Situation assessment of an autonomous emergency
brake for arbitrary vehicle-to-vehicle collision scenar-
ios. IEEE Transactions on Intelligent Transportation
Systems, 10(4):678–687.
Kingma, D. P. and Ba, J. (2014). Adam: A method for
stochastic optimization.
Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T.,
Tassa, Y., Silver, D., and Wierstra, D. (2015). Continu-
ous control with deep reinforcement learning.
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A.,
Antonoglou, I., Wierstra, D., and Riedmiller, M. (2013).
Playing atari with deep reinforcement learning.
Porav, H. and Newman, P. (2018). Imminent collision miti-
gation with reinforcement learning and vision. In 2018
21st International Conference on Intelligent Trans-
portation Systems (ITSC), pages 958–964.
Sutton, R. S. and Barto, A. G. (2018). Reinforcement Learn-
ing: An Introduction. A Bradford Book, Cambridge,
MA, USA.
Sutton, R. S., McAllester, D. A., Singh, S. P., and Mansour,
Y. (2000). Policy gradient methods for reinforcement
learning with function approximation. In Advances
in neural information processing systems, pages 1057–
1063.
Talpaert., V., Sobh., I., Kiran., B. R., Mannion., P., Yoga-
mani., S., El-Sallab., A., and Perez., P. (2019). Ex-
ploring applications of deep reinforcement learning
for real-world autonomous driving systems. In Pro-
ceedings of the 14th International Joint Conference
on Computer Vision, Imaging and Computer Graphics
Theory and Applications - Volume 5: VISAPP,, pages
564–572. INSTICC, SciTePress.
Uhlenbeck, G. E. and Ornstein, L. S. (1930). On the theory
of the brownian motion. Phys. Rev., 36:823–841.
Vasquez, R. and Farooq, B. (2019). Multi-objective au-
tonomous braking system using naturalistic dataset. In
2019 IEEE Intelligent Transportation Systems Confer-
ence (ITSC), pages 4348–4353.
Watkins, C. J. and Dayan, P. (1992). Q-learning. Machine
learning, 8(3-4):279–292.
ICAART 2021 - 13th International Conference on Agents and Artificial Intelligence
180