they used. We have also detailed how these
architectures vary in the use of attention mechanisms.
We have found that CAML, Text-CNN and BERT are
the most commonly used base architectures, while
label-wise and self-attention are the most common
mechanisms to develop such novel
architectures. Many of these studies build on
architectures from previous similar studies to
introduce a pipeline with better results in the area. We
can also see the popular use of attention mechanisms
with transformer models, CNNs and even RNNs.
However, all of these researches are focused on
evaluation of label predictions and visualization of
attention scores. In order to develop plausible
interpretability, a standard set of explainability
evaluations should be performed. We recognize an
immediate need to not only consider explainability
that is much more relatable to end-users’ cognition
and learning, but is also objective in terms of being
measurable and comparable. Such evaluation
measures combined with overall performance of the
NLP and AI models in the healthcare domain would
aid the readers as well as the end-users in
understanding the outputs better, hence working with
these AI models and systems with higher confidence
and trust.
REFERENCES
Biswas, B., Pham, T. H., & Zhang, P. (2021, June).
Transicd: Transformer based code-wise attention
model for explainable icd coding. In International
Conference on Artificial Intelligence in Medicine (pp.
469-478). Springer, Cham.
Cabitza, F., Campagner, A., & Datteri, E. (2021). To err is
(only) human. Reflections on how to move from
accuracy to trust for medical AI. In Exploring
Innovation in a Digital World (pp. 36-49). Springer,
Cham.
Castelvecchi, D. (2016). Can we open the black box of AI?.
Nature News, 538(7623), 20.
DeGrave, A. J., Janizek, J. D., & Lee, S. I. (2021). AI for
radiographic COVID-19 detection selects shortcuts
over signal. Nature Machine Intelligence, 3(7), 610-619
Dong, H., Suárez-Paniagua, V., Whiteley, W., & Wu, H.
(2021). Explainable automated coding of clinical notes
using hierarchical label-wise attention networks and
label embedding initialisation. Journal of biomedical
informatics, 116, 103728.
Feng, J., Shaib, C., & Rudzicz, F. (2020, November).
Explainable clinical decision support from text. In
Proceedings of the 2020 conference on empirical
methods in natural language processing (EMNLP) (pp.
1478-1489).
Feucht, M., Wu, Z., Althammer, S., & Tresp, V. (2021).
Description-based Label Attention Classifier for
Explainable ICD-9 Classification. arXiv preprint
arXiv:2109.12026.
IBM. Explainable AI (XAI). https://www.ibm.com/
watson/explainable-ai. Accessed [Nov, 2022]
Hu, D. (2019, September). An introductory survey on
attention mechanisms in NLP problems. In Proceedings
of SAI Intelligent Systems Conference (pp. 432-448).
Springer, Cham.
Hu, S., Teng, F., Huang, L., Yan, J., & Zhang, H. (2021).
An explainable CNN approach for medical codes
prediction from clinical text. BMC Medical Informatics
and Decision Making, 21(9), 1-12.
Hunter, K. (1998). ScienceDirect™. The Serials Librarian,
33(3-4), 287-297.
Johnson, A. E., Pollard, T. J., Shen, L., Lehman, L. W. H.,
Feng, M., Ghassemi, M., ... & Mark, R. G. (2016).
MIMIC-III, a freely accessible critical care database.
Scientific data, 3(1), 1-9.
Li, X., Zhang, Y., Dong, D., Wei, H., & Lu, M. (2021).
JLAN: medical code prediction via joint learning
attention networks and denoising mechanism. BMC
bioinformatics, 22(1), 1-21.
Liu, L., Perez-Concha, O., Nguyen, A., Bennett, V., &
Jorm, L. (2022). Hierarchical Label-wise Attention
Transformer Model for Explainable ICD Coding. arXiv
preprint arXiv:2204.10716.
Mayya, V., Kamath, S. S., & Sugumaran, V. (2021,
October). Label Attention Transformer Architectures
for ICD-10 Coding of Unstructured Clinical Notes. In
2021 IEEE Conference on Computational Intelligence
in Bioinformatics and Computational Biology (CIBCB)
(pp. 1-7). IEEE.
Mayya, V., Kamath, S., Krishnan, G. S., & Gangavarapu,
T. (2021). Multi-channel, convolutional attention based
neural model for automated diagnostic coding of
unstructured patient discharge summaries. Future
Generation Computer Systems,
118, 374-391.
Mullenbach, J., Wiegreffe, S., Duke, J., Sun, J., &
Eisenstein, J. (2018). Explainable prediction of medical
codes from clinical text. arXiv preprint
arXiv:1802.05695.
Naylor, M., French, C., Terker, S., & Kamath, U. (2021).
Quantifying explainability in nlp and analyzing
algorithms for performance-explainability tradeoff.
arXiv preprint arXiv:2107.05693.
Nguyen, B., & Ji, S. (2021). Fine-tuning Pretrained
Language Models with Label Attention for Explainable
Biomedical Text Classification. arXiv preprint
arXiv:2108.11809.
Nourani, M., King, J., & Ragan, E. (2020, October). The
role of domain expertise in user trust and the impact of
first impressions with intelligent systems. In
Proceedings of the AAAI Conference on Human
Computation and Crowdsourcing (Vol. 8, pp. 112-121).
Savage, N. (2022). Breaking into the black box of artificial
intelligence. Nature, March 2022 https://doi.org/10.10
38/d41586-022-00858-1