
the proposed method, has more than twice as much
concept detectors in the category scene than the basis
suggested by UIBE. Additionally its concept detec-
tors are able to detect 25 scene concepts more than
the detectors of UIBE. For a network that is trained
to do scene classification this suggestion looks highly
plausible. An even more prominent result regards
ResNet50. Compared to UIBE, the proposed method
suggests a basis with 10 times more concept detec-
tors, for concepts in the action category. Addition-
ally, the respective basis’ concept detectors can detect
47 more action concepts than the detectors of UIBE.
This suggestion aligns better with the goal of a net-
work that is trained to perform action recognition.
5 CONCLUSION
In this work we proposed to complement previous
work (UIBE) with a novel loss term, that exploits
the knowledge encoded in CNN image classifiers and
suggests more interpretable bases. The proposed
method demonstrates up to 45.8% interpretability im-
provements in the extracted bases, when using opti-
mal hyper-parameters that were suggested for learn-
ing a basis regarding a different classifier trained on
another task. Future work may study applications of
the proposed method to debug and improve model
performance.
ACKNOWLEDGEMENT
This work has been supported by the EC funded Hori-
zon Europe Framework Programme: CAVAA Grant
Agreement 101071178.
REFERENCES
Achtibat, R., Dreyer, M., Eisenbraun, I., Bosse, S., Wie-
gand, T., Samek, W., and Lapuschkin, S. (2023). From
attribution maps to human-understandable explana-
tions through concept relevance propagation. Nature
Machine Intelligence, 5(9).
Alain, G. and Bengio, Y. (2016). Understanding
intermediate layers using linear classifier probes.
arXiv:1610.01644.
Bach, S., Binder, A., Montavon, G., Klauschen, F., M
¨
uller,
K.-R., and Samek, W. (2015). On pixel-wise explana-
tions for non-linear classifier decisions by layer-wise
relevance propagation. PloS one, 10(7).
Bau, D., Zhou, B., Khosla, A., Oliva, A., and Torralba,
A. (2017). Network dissection: Quantifying inter-
pretability of deep visual representations. In CVPR.
Chormai, P., Herrmann, J., M
¨
uller, K.-R., and Montavon, G.
(2022). Disentangled explanations of neural network
predictions by finding relevant subspaces.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-
Fei, L. (2009). Imagenet: A large-scale hierarchical
image database. In CVPR. IEEE.
Doumanoglou, A., Asteriadis, S., and Zarpalas, D.
(2023). Unsupervised interpretable basis extraction
for concept-based visual explanations. Acccepted
at IEEE TAI, currently available on arXiv preprint
arXiv:2303.10523.
Fong, R. and Vedaldi, A. (2018). Net2vec: Quantifying and
explaining how concepts are encoded by filters in deep
neural networks. In CVPR.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In CVPR.
Kim, B., Wattenberg, M., Gilmer, J., Cai, C., Wexler, J.,
Viegas, F., et al. (2018). Interpretability beyond fea-
ture attribution: Quantitative testing with concept ac-
tivation vectors (TCAV). In ICML. PMLR.
Losch, M., Fritz, M., and Schiele, B. (2021). Semantic bot-
tlenecks: Quantifying and improving inspectability of
deep representations. IJCV, 129(11).
Monfort, M., Andonian, A., Zhou, B., Ramakrishnan, K.,
Bargal, S. A., Yan, T., Brown, L., Fan, Q., Gutfre-
und, D., Vondrick, C., et al. (2019). Moments in time
dataset: one million videos for event understanding.
TPAMI, 42(2).
Mu, J. and Andreas, J. (2020). Compositional explanations
of neurons. In NIPS, volume 33. Curran Associates,
Inc.
Ramakrishnan, K., Monfort, M., McNamara, B. A., Las-
celles, A., Gutfreund, D., Feris, R. S., and Oliva, A.
(2019). Identifying interpretable action concepts in
deep networks. In CVPR Workshops.
Rapin, J. and Teytaud, O. (2018). Nevergrad - A
gradient-free optimization platform. https://GitHub.
com/FacebookResearch/Nevergrad.
Simonyan, K. and Zisserman, A. (2015). Very deep convo-
lutional networks for large-scale image recognition. In
Bengio, Y. and LeCun, Y., editors, ICLR.
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.,
Anguelov, D., Erhan, D., Vanhoucke, V., and Rabi-
novich, A. (2015). Going deeper with convolutions.
In CVPR.
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Er-
han, D., Goodfellow, I., and Fergus, R. (2013). In-
triguing properties of neural networks. arXiv preprint
arXiv:1312.6199.
Vielhaben, J., Bl
¨
ucher, S., and Strodthoff, N. (2022). Sparse
subspace clustering for concept discovery (SSCCD).
Zarlenga, M. E., Barbiero, P., Shams, Z., Kazhdan, D.,
Bhatt, U., Weller, A., and Jamnik, M. (2023). Towards
robust metrics for concept representation evaluation.
arXiv preprint arXiv:2301.10367.
Zhang, R., Madumal, P., Miller, T., Ehinger, K. A., and Ru-
binstein, B. I. P. (2021). Invertible concept-based ex-
planations for cnn models with non-negative concept
activation vectors. In AAAI.
Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., and Tor-
ralba, A. (2017). Places: A 10 million image database
for scene recognition. IEEE TPAMI.
Zhou, B., Sun, Y., Bau, D., and Torralba, A. (2018). Inter-
pretable basis decomposition for visual explanation.
In ECCV.
VISAPP 2024 - 19th International Conference on Computer Vision Theory and Applications
424