cases with very basic classification models. Our ap-
proach works in a light-weight and post-processing
manner, i.e., we do not modify the model nor need
knowledge of the process used by the attacker for gen-
erating adversarial examples. We achieve averaged
detection accuracy values of up to 100% for differ-
ent network architectures and datasets. Moreover, it
has to be pointed out, that our proposed detection ap-
proach is the first that was not designed for a specific
adversarial attack, but has a high detection capability
across multiple types. Given the high detection accu-
racy and the simplicity of the proposed approach, we
are convinced, that it should serve as simple baseline
for more elaborated but computationally more expen-
sive approaches developed in future.
ACKNOWLEDGEMENTS
This work is supported by the Ministry of Culture
and Science of the German state of North Rhine-
Westphalia as part of the KI-Starter research fund-
ing program and by the Deutsche Forschungsgemein-
schaft (DFG, German Research Foundation) under
Germany’s Excellence Strategy – EXC-2092 CASA
– 390781972.
REFERENCES
Agnihotri, S. and Keuper, M. (2023). Cospgd: a unified
white-box adversarial attack for pixel-wise prediction
tasks. 2, 4
Arnab, A., Miksik, O., and Torr, P. (2018). On the robust-
ness of semantic segmentation models to adversarial
attacks. In IEEE/CVF Conference on Computer Vi-
sion and Pattern Recognition (CVPR). 2, 3, 6
Bar, A., Lohdefink, J., Kapoor, N., Varghese, S., Huger, F.,
Schlicht, P., and Fingscheidt, T. (2021). The vulner-
ability of semantic segmentation networks to adver-
sarial attacks in autonomous driving: Enhancing ex-
tensive environment sensing. IEEE Signal Processing
Magazine. 1, 3, 6
Bryniarski, O., Hingun, N., Pachuca, P., Wang, V., and Car-
lini, N. (2022). Evading adversarial example detection
defenses with orthogonal projected gradient descent.
In International Conference on Learning Representa-
tions (ICLR). 4
Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., and
Adam, H. (2018). Encoder-decoder with atrous sep-
arable convolution for semantic image segmentation.
In European Conference on Computer Vision (ECCV).
1, 2, 5
Cho, S., Jun, T. J., Oh, B., and Kim, D. (2020). Dapas
: Denoising autoencoder to prevent adversarial attack
in semantic segmentation. In International Joint Con-
ference on Neural Network (IJCNN). 3
Chollet, F. (2017). Xception: Deep learning with depthwise
separable convolutions. IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR). 5
Cisse, M., Adi, Y., Neverova, N., and Keshet, J. (2017).
Houdini: Fooling deep structured prediction models.
In Conference on Neural Information Processing Sys-
tems (NeurIPS). 2, 4
Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler,
M., Benenson, R., Franke, U., Roth, S., and Schiele,
B. (2016). The cityscapes dataset for semantic urban
scene understanding. In IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR). 2, 5
Everingham, M., Van Gool, L., Williams, C. K. I.,
Winn, J., and Zisserman, A. (2012). The
PASCAL Visual Object Classes Challenge
2012 (VOC2012) Results. http://www.pascal-
network.org/challenges/VOC/voc2012/workshop/
index.html. 2, 5
Feinman, R., Curtin, R. R., Shintre, S., and Gardner, A. B.
(2017). Detecting adversarial samples from artifacts.
2
Goodfellow, I. J., Shlens, J., and Szegedy, C. (2015). Ex-
plaining and harnessing adversarial examples. In Ben-
gio, Y. and LeCun, Y., editors, International Confer-
ence on Learning Representations (ICLR). 2, 3
Gu, J., Zhao, H., Tresp, V., and Torr, P. (2022). Segpgd: An
effective and efficient adversarial attack for evaluating
and boosting segmentation robustness. In European
Conference on Computer Vision (ECCV). 2, 4
He, X., Yang, S., Li, G., Li, H., Chang, H., and Yu, Y.
(2019). Non-local context encoder: Robust biomedi-
cal image segmentation against adversarial attacks. In
AAAI Conference on Artificial Intelligence. 3, 6
Hendrycks, D. and Gimpel, K. (2016). A baseline for de-
tecting misclassified and out-of-distribution examples
in neural networks. 3
Khamaiseh, S. Y., Bagagem, D., Al-Alaj, A., Mancino,
M., and Alomari, H. W. (2022). Adversarial deep
learning: A survey on adversarial attacks and defense
mechanisms on image classification. IEEE Access. 2
Klingner, M., B
¨
ar, A., and Fingscheidt, T. (2020). Im-
proved noise and attack robustness for semantic seg-
mentation by using multi-task training with self-
supervised depth estimation. In IEEE/CVF Confer-
ence on Computer Vision and Pattern Recognition
Workshop (CVPRW). 2, 3, 6
Krizhevsky, A. (2009). Learning multiple layers of features
from tiny images. 2
Kurakin, A., Goodfellow, I. J., and Bengio, S. (2017). Ad-
versarial machine learning at scale. In International
Conference on Learning Representations (ICLR). 2,
3, 6
LeCun, Y. and Cortes, C. (2010). MNIST handwritten digit
database. 2
Maag, K., Chan, R., Uhlemeyer, S., Kowol, K., and
Gottschalk, H. (2022). Two video data sets for track-
ing and retrieval of out of distribution objects. In Asian
Conference on Computer Vision (ACCV), pages 3776–
3794. 3
Uncertainty-Based Detection of Adversarial Attacks in Semantic Segmentation
45