The double lane change (ISO 3888-1) is a pure
validation maneuver and thus evokes a driving
dynamic that was not explicitly trained by the agent.
However, this driving dynamic is comparable to that
of a slalom, so that similar results can be expected. It
is validated exemplary at 60 km/h.
As expected, the control behaviour of the neural
controller is very similar to that of slalom driving. It
shows that the control behaviour can also be
extrapolated to other maneuvers and driving
situations.
8 CONCLUSIONS
In the context of this work an active roll control with
an artificial neural network based on an actor-critic
reinforcement learning method has been successfully
realized. The neural controller was realized with the
TensorFlow libraries in a Python script and combined
with the simulation model of the entire vehicle and
the active roll stabilization contained therein via a
TCP/IP interface.
A guaranteed calculation of the torque to be set in
a fixed time interval and a time limit of the waiting
time of the TCP/IP interface created a real-time
control. If, in the defined waiting time, the actuator
does not receive any action from the neural network,
the torque is set from the previous time step. The
developed neural controller is able, at any time, to
stably reduce the roll angle caused by the centrifugal
force of the vehicle body by means of an actuator. The
functionality of the controller is thus given.
The results show that the developed controller
produces a rather uneven roll behaviour for both
directions of the steering angle in comparison to
established, conventional controllers. However, it has
been proven that roll stabilization by artificial neural
networks is possible and that the developed model is
able to replace conventional controllers. If the
knowledge gained in this work continues to be
applied to the model and extended with small and
precise optimizations, a neural controller with
symmetric behaviour can be trained for lateral
acceleration in both directions. Since the field of
machine learning works with very complex contexts
and is strongly randomized, this is a matter of time.
Basically, in 100 training runs with identical
hyperparameters, 100 different results can be
achieved, the extent of which is far from expedient.
Nonetheless, it has been shown that the neural
network used can provide a controller with tolerable
results. A fixed reproducibility of this result is not
given by the immense influence of randomness, but
due to the stochastics also better results are possible.
Due to its structure, the agent is able to adjust its
weights so that, for positive lateral accelerations, an
at least equal reduction of the roll angle is achieved,
as for negative transverse accelerations.
Further works will investigate the influence and
possible improvements by applying a regularization
on the weight adjustment to ensure the minimal
optimal weights and symmetric behaviour for
positive and negative lateral accelerations.
REFERENCES
Boada, M., Boada, B., Gauchia Babe, A. Calvo, J. and Diaz,
V., 2009. Active roll control using reinforcement
learning for a single unit heavy vehicle. In International
Journal of Heavy Vehicle Systems, 16(4), pp. 412-430.
Fu, Z.-J., Li, B., Ning, X.-B., Xie and W.-D., 2017. Online
Adaptive Optimal Control of Vehicle Active
Suspension Systems Using Single-Network
Approximate Dynamic Programming. In Mathematical
Problems in Engineering.
Goodfellow, I., Bengio, Y. and Courville, A., 2016. Deep
Learning, The MIT Press. London.
Hebb, D., 1949. The Organization of Behavior, John Wiley
& Sons, Inc. New York.
IPG, 2019. IPG Automotive’s Official Website. [online]
Available at https://ipg-automotive.com/ [Accessed
09 Feb. 2019].
Karlik, B. and Olgac, V., 2011. Performance Analysis of
Various Activation Functions in Generalized MLP
Architectures of Neural Networks. In International
Journal of Artificial Intelligence and Expert Systems
(IJAE), 1(4), pp. 111-122.
Konda, V. and Tsitsiklis, J., 2003. Actor-Critic Algorithms.
In SIAM Journal on Control and Optimization, 42(4),
pp. 1143-1166.
Li, H., Guo, C. and Jin, H., 2005. Design of Adaptive
Inverse Mode Wavelet Neural Network Controller of
Fin Stabilizer. In International Conference on Neural
Networks and Brain, 2005, 3, pp. 1745-1748.
MathWorks, 2019. MathWorks’ Official Website. [online]
Available at https://de.mathworks.com/ [Accessed 09
Feb. 2019].
Rosenblatt, F., 1958. The Perceptron: A probabilistic model
for information storage and organization in the brain. In
Psychological Review, 65(6), pp. 386-408.
Schramm, D., Hiller, M. and Bardini, R., 2018. Vehicle
Dynamics – Modeling and Simulation, Springer.
2
nd
Edition.
Sieberg, P., Reicherts, S. and Schramm, D., 2018.
Nichtlineare modellbasierte prädiktive Regelung zur
aktiven Wankstabilisierung von Personenkraftwagen.
In 4
th
IFToMM D-A-CH Konferenz 2018.
Sutton, R. and Barto, A., 2018. Reinforcement Learning: An
Introduction, The MIT Press. London, 2
nd
edition.