
unwavering support and dedication have been instru-
mental in the success of this research. A special note
of thanks is extended to the medical coders who have
shown exceptional commitment and diligence. Their
tireless efforts and invaluable support have signifi-
cantly enriched our work.
This work has been achieved in the frame of the
EIPHI Graduate school (contract ”ANR-17-EURE-
0002”).
CONFLICT OF INTEREST
The authors declare that they have no known com-
peting financial interests or personal relationships that
could have appeared to influence the work reported in
this paper.
REFERENCES
Al-Anzi, F. and AbuZeina, D. (2020). Enhanced latent se-
mantic indexing using cosine similarity measures for
medical application. The International Arab Journal
of Information Technology, 17(5):742–749.
Al-Bashabsheh, E., Alaiad, A., Al-Ayyoub, M., Beni-Yonis,
O., Zitar, R. A., and Abualigah, L. (2023). Improving
clinical documentation: automatic inference of icd-10
codes from patient notes using bert model. The Jour-
nal of Supercomputing, 79(11):12766–12790.
Alsentzer, E., Murphy, J., Boag, W., Weng, W.-H., Jin, D.,
Naumann, T., and McDermott, M. (2019). Publicly
available clinical BERT embeddings. In Proceed-
ings of the 2nd Clinical Natural Language Process-
ing Workshop, pages 72–78. Association for Compu-
tational Linguistics.
Boldini, D., Friedrich, L., Kuhn, D., and Sieber, S. A.
(2022). Tuning gradient boosting for imbalanced
bioassay modelling with custom loss functions. Jour-
nal of Cheminformatics, 14(1).
del Barrio, E., Gordaliza, P., and Loubes, J.-M. (2020). Re-
view of mathematical frameworks for fairness in ma-
chine learning. ArXiv.
Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2019).
BERT: pre-training of deep bidirectional transformers
for language understanding. In NAACL-HLT 2019,
Minneapolis, MN, USA, June 2-7, Volume 1, pages
4171–4186. Association for Computational Linguis-
tics.
Dinkel, H., Wu, M., and Yu, K. (2019). Text-based depres-
sion detection on sparse data. arXiv.
Esteva, A., Robicquet, A., Ramsundar, B., Kuleshov, V.,
DePristo, M., Chou, K., Cui, C., Corrado, G., Thrun,
S., and Dean, J. (2019). A guide to deep learning in
healthcare. Nature Medicine, 25(1):24–29.
Ferger, W. F. (1931). The nature and use of the harmonic
mean. Journal of the American Statistical Association,
26(173):36–40.
Giyahchi, T., Singh, S., Harris, I., and Pechmann, C. (2022).
Customized training of pretrained language models to
detect post intents in online health support groups.
Multimodal AI in Healthcare, pages 59–75.
Grabner, C., Safont-Andreu, A., Burmer, C., and Schekoti-
hin, K. (2022). A bert-based report classification for
semiconductor failure analysis. International Sympo-
sium for Testing and Failure Analysis.
Hatoum, M., Charr, J.-C., Guyeux, C., Laiymani, D., and
Ghaddar, A. (2023). Emte: An enhanced medical
terms extractor using pattern matching rules. 15th In-
ternational Conference on Agents and Artificial Intel-
ligence, pages 301–311.
Hatoum, M. B., Charr, J. C., Ghaddar, A., Guyeux, C., and
Laiymani, D. (2024a). Nnbsvr: Neural network-based
semantic vector representations of icd-10 codes. Un-
der revision.
Hatoum, M. B., Charr, J. C., Ghaddar, A., Guyeux, C.,
and Laiymani, D. (2024b). Utp: A unified term pre-
sentation tool for clinical textual data using pattern-
matching rules and dictionary-based ontologies. Lec-
ture Notes in Computer Science, pages 353–369.
Kulkarni, D., Ghosh, A., Girdhari, A., Liu, S., Vance,
L. A., Unruh, M., and Sarkar, J. (2024). Enhancing
pre-trained contextual embeddings with triplet loss as
an effective fine-tuning method for extracting clinical
features from electronic health record derived mental
health clinical notes. Natural Language Processing
Journal, 6:100045.
Kumari, S. and Pushphavati, T. (2022). Question answer-
ing and text generation using bert and gpt-2 model. In
Computational Methods and Data Engineering: Pro-
ceedings of ICCMDE 2021, pages 93–110. Springer.
Kurokawa, R., Ohizumi, Y., Kanzawa, J., Kurokawa, M.,
Kiguchi, T., Gonoi, W., and Abe, O. (2024). Diagnos-
tic performance of claude 3 from patient history and
key images in diagnosis please cases. medRxiv.
Long, R. (2021). Fairness in machine learning: Against
false positive rate equality as a measure of fairness.
Journal of Moral Philosophy, 19(1):49–78.
Mihalache, A., Grad, J., Patil, N. S., Huang, R. S., Popovic,
M. M., Mallipatna, A., Kertes, P. J., and Muni, R. H.
(2024). Google gemini and bard artificial intelligence
chatbot performance in ophthalmology knowledge as-
sessment. Eye, pages 2530–2535.
Mittelstadt, B., Wachter, S., and Russell, C. (2023). The un-
fairness of fair machine learning: Levelling down and
strict egalitarianism by default. Michigan Technology
Law Review.
Mohammadi, S. and Chapon, M. (2020). Investigating the
performance of fine-tuned text classification models
based-on bert. In 2020 IEEE 22nd International Con-
ference on High Performance Computing and Com-
munications, pages 1252–1257.
Mou, C., Ye, X., Wu, J., and Dai, W. (2023). Automated
icd coding based on neural machine translation. In
2023 8th International Conference on Cloud Comput-
ing and Big Data Analytics (ICCCBDA), pages 495–
500. IEEE.
Beyond Equality Matching: Custom Loss Functions for Semantics-Aware ICD-10 Coding
173