GPU (32GB), where the CPU-memory consumption
was roughly 90GB, due to the in memory loaded
dataset of images. InfoSalGAIL framework is based
on TensorFlow (version 1.15) and Keras 2.0 library,
where the connection to OpenDS has been realized
though a transmission control protocol (TCP) to guar-
antee no data loss.
5 CONCLUSIONS
In this paper, we presented a novel approach, named
InfoSalGAIL, for visual attention-empowered imita-
tion learning of pedestrian behavior in critical traffic
scenarios that can handle substantial state-space and
variations, e.g. on-street urban scenarios, to mimic
complex human-like behavior of experts in a virtual
environment. Moreover, we synthesised two classes
of navigation (cf. in Section 3) which renders InfoSal-
GAIL quite suitable for the challenge of critical traffic
scenario generation. Our experimets revealed that In-
foSalGAIL can significantly outperform the selected
baseline InfoGAIL for the given objective due to the
utilization of a saliency map and its direct impact
on the policy generator in deriving the output vector
(control actions). To support this research activity, the
functionality of the OpenDS simulation software has
been extended to allow for a pedestrian-centric con-
trol, resulting in a creation of a new dataset, which
consists of more than 140K pairs of images and cor-
responding saliency maps generated from a virtual
clone of Saarbruecken city (Germany).
Future research is concerned with an extension of
the saliency generator network by incorporating latent
variables to further differentiate between the pedes-
trian imitating avatars such in terms of age, average
speed, short term interests. In this regard, the created
benchmark will be extended with a new set of realistic
scenarios based on JAAD
5
dataset to capture ground
truth data. In general, we hope that InfoSalGAIL at-
tracts more attention to the topic of human-like behav-
ior simulation in the scope of generating critical traffic
scenarios for virtual tests and validation of collision-
free navigation methods of self-driving cars.
ACKNOWLEDGEMENTS
This research was funded by the German Federal
Ministry for Education and Research (BMBF) in the
project REACT under grant 01IW17003.
5
http://data.nvision2.eecs.yorku.ca/JAAD\ dataset
REFERENCES
Abbeel, P. and Ng, A. Y. (2004). Apprenticeship learning
via inverse reinforcement learning. In Proceedings of
the twenty-first international conference on Machine
learning, page 1.
Arjovsky, M., Chintala, S., and Bottou, L. (2017). Wasser-
stein gan. arXiv preprint arXiv:1701.07875.
Bloem, M. and Bambos, N. (2014). Infinite time horizon
maximum causal entropy inverse reinforcement learn-
ing. In 53rd IEEE Conference on Decision and Con-
trol, pages 4911–4916. IEEE.
Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever,
I., and Abbeel, P. (2016). Infogan: Interpretable rep-
resentation learning by information maximizing gen-
erative adversarial nets. In Advances in neural infor-
mation processing systems, pages 2172–2180.
Corbetta, M. (1998). Frontoparietal cortical networks for
directing attention and the eye to visual locations:
identical, independent, or overlapping neural sys-
tems? Proceedings of the National Academy of Sci-
ences, 95(3):831–838.
Duchi, J., Hazan, E., and Singer, Y. (2011). Adaptive sub-
gradient methods for online learning and stochastic
optimization. Journal of machine learning research,
12(7).
Englert, P. and Toussaint, M. (2018). Inverse kkt–learning
cost functions of manipulation tasks from demonstra-
tions. In Robotics Research, pages 57–72. Springer.
Ermon, S., Xue, Y., Toth, R., Dilkina, B., Bernstein, R.,
Damoulas, T., Clark, P., DeGloria, S., Mude, A.,
Barrett, C., et al. (2015). Learning large-scale dy-
namic discrete choice models of spatio-temporal pref-
erences with application to migratory pastoralism in
east africa. In Twenty-Ninth AAAI Conference on Ar-
tificial Intelligence.
Finn, C., Levine, S., and Abbeel, P. (2016). Guided cost
learning: Deep inverse optimal control via policy op-
timization. In International conference on machine
learning, pages 49–58.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B.,
Warde-Farley, D., Ozair, S., Courville, A., and Ben-
gio, Y. (2014). Generative adversarial nets. In
Advances in neural information processing systems,
pages 2672–2680.
Hausman, K., Chebotar, Y., Schaal, S., Sukhatme, G., and
Lim, J. J. (2017). Multi-modal imitation learning from
unstructured demonstrations using generative adver-
sarial nets. In Advances in Neural Information Pro-
cessing Systems, pages 1235–1245.
Ho, J. and Ermon, S. (2016). Generative adversarial im-
itation learning. In Advances in neural information
processing systems, pages 4565–4573.
Ho, J., Gupta, J., and Ermon, S. (2016). Model-free imi-
tation learning with policy optimization. In Interna-
tional Conference on Machine Learning, pages 2760–
2769.
Jolliffe, I. (2003). Principal component analysis. Techno-
metrics, 45(3):276.
NCTA 2020 - 12th International Conference on Neural Computation Theory and Applications
334