
drop in confidence when presented with very noisy
inputs.
Our work, despite being the first analysis on the
uncertainty of the NNs when directly modeling in-
put uncertainty, is still quite small scale and mostly
observational, and could potentially benefit for more
extensive analyses. For instance, larger-scale datasets
might be used—although BNNs are notoriously dif-
ficult and slow to train on bigger datasets. Also, we
could extend the selection of BNN-training schemes
to other methods, like the more recent SWAG (Mad-
dox et al., 2019), or Hamiltonian Monte–Carlo (Neal
et al., 2011), which is still considered the golden stan-
dard for Bayesian modeling, albeit very unfeasible to
apply in the large-scale datasets used in modern Deep
Learning. Finally, our study could benefit from the
addition on the analysis of uncertainty disentangle-
ment by the BNNs, to understand to what extent the
models are able to integrate aleatoric uncertainty into
the input uncertainty.
REFERENCES
Blundell, C., Cornebise, J., Kavukcuoglu, K., and Wier-
stra, D. (2015). Weight uncertainty in neural network.
In Bach, F. and Blei, D., editors, Proceedings of the
32nd International Conference on Machine Learning,
volume 37 of Proceedings of Machine Learning Re-
search, pages 1613–1622, Lille, France. PMLR.
Drucker, H. and Le Cun, Y. (1992). Improving gener-
alization performance using double backpropagation.
IEEE transactions on neural networks, 3(6):991–997.
Gal, Y. and Ghahramani, Z. (2016). Dropout as a bayesian
approximation: Representing model uncertainty in
deep learning. In international conference on machine
learning, pages 1050–1059. PMLR.
Guo, C., Pleiss, G., Sun, Y., and Weinberger, K. Q. (2017).
On calibration of modern neural networks. In Interna-
tional conference on machine learning, pages 1321–
1330. PMLR.
Harris, E., Marcu, A., Painter, M., Niranjan, M., Pr
¨
ugel-
Bennett, A., and Hare, J. (2020). Fmix: Enhanc-
ing mixed sample data augmentation. arXiv preprint
arXiv:2002.12047.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Identity
mappings in deep residual networks. In Computer
Vision–ECCV 2016: 14th European Conference, Am-
sterdam, The Netherlands, October 11–14, 2016, Pro-
ceedings, Part IV 14, pages 630–645. Springer.
Hinton, G. E., Srivastava, N., Krizhevsky, A., Sutskever, I.,
and Salakhutdinov, R. R. (2012). Improving neural
networks by preventing co-adaptation of feature de-
tectors. arXiv preprint arXiv:1207.0580.
H
¨
ullermeier, E. (2014). Learning from imprecise and fuzzy
observations: Data disambiguation through general-
ized loss minimization. International Journal of Ap-
proximate Reasoning, 55(7):1519–1534.
H
¨
ullermeier, E. and Waegeman, W. (2021). Aleatoric and
epistemic uncertainty in machine learning: An intro-
duction to concepts and methods. Machine learning,
110(3):457–506.
Kingma, D. P. and Ba, J. (2014). Adam: A
method for stochastic optimization. arXiv preprint
arXiv:1412.6980.
Kramer, O. and Kramer, O. (2016). Scikit-learn. Machine
learning for evolution strategies, pages 45–53.
Lakshminarayanan, B., Pritzel, A., and Blundell, C. (2017).
Simple and scalable predictive uncertainty estimation
using deep ensembles. Advances in neural informa-
tion processing systems, 30.
LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998).
Gradient-based learning applied to document recogni-
tion. Proceedings of the IEEE, 86(11):2278–2324.
MacKay, D. J. C. (1992). The evidence framework ap-
plied to classification networks. Neural Computation,
4(5):720–736.
Maddox, W. J., Izmailov, P., Garipov, T., Vetrov, D. P., and
Wilson, A. G. (2019). A simple baseline for bayesian
uncertainty in deep learning. Advances in neural in-
formation processing systems, 32.
Naeini, M. P., Cooper, G., and Hauskrecht, M. (2015).
Obtaining well calibrated probabilities using bayesian
binning. In Proceedings of the AAAI conference on
artificial intelligence, volume 29.
Neal, R. M. et al. (2011). Mcmc using hamiltonian dynam-
ics. Handbook of markov chain monte carlo, 2(11):2.
Nguyen, A., Yosinski, J., and Clune, J. (2015). Deep neural
networks are easily fooled: High confidence predic-
tions for unrecognizable images. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 427–436.
Nix, D. A. and Weigend, A. S. (1994). Estimating the mean
and variance of the target probability distribution. In
Proceedings of 1994 ieee international conference on
neural networks (ICNN’94), volume 1, pages 55–60.
IEEE.
Ovadia, Y., Fertig, E., Ren, J., Nado, Z., Sculley, D.,
Nowozin, S., Dillon, J., Lakshminarayanan, B., and
Snoek, J. (2019). Can you trust your model's uncer-
tainty? evaluating predictive uncertainty under dataset
shift. In Wallach, H., Larochelle, H., Beygelzimer,
A., d'Alch
´
e-Buc, F., Fox, E., and Garnett, R., editors,
Advances in Neural Information Processing Systems,
volume 32. Curran Associates, Inc.
Papadopoulos, G., Edwards, P. J., and Murray, A. F. (2001).
Confidence estimation methods for neural networks:
A practical comparison. IEEE transactions on neural
networks, 12(6):1278–1287.
Roberts, H. V. (1965). Probabilistic prediction. Journal of
the American Statistical Association, 60(309):50–62.
Rodrigues, N. V., Abramo, L. R., and Hirata, N. S. (2023).
The information of attribute uncertainties: what con-
volutional neural networks can learn about errors in
input data. Machine Learning: Science and Technol-
ogy, 4(4):045019.
Tzelepis, C., Mezaris, V., and Patras, I. (2017). Linear
maximum margin classifier for learning from uncer-
VISAPP 2025 - 20th International Conference on Computer Vision Theory and Applications
198