
cantly higher GPU resources for fine-tuning and in-
ference. Notably, ChatGPT 4 underperformed in this
task, highlighting the limitations of general-purpose
LLMs without domain-specific fine-tuning.
In terms of computational efficiency, the Boosted
Bi-GRU model demonstrated the best trade-off be-
tween accuracy and resource usage, while models like
Phi and BiomedLM provided a balance of scalabil-
ity and performance in biomedical contexts. These
findings underscore the importance of aligning model
selection and fine-tuning strategies with task-specific
requirements and resource constraints.
Future work will explore advanced parameter-
efficient fine-tuning techniques, such as adapters or
LoRA, to further enhance the capabilities of large
models while minimizing computational costs. Addi-
tionally, integrating more sophisticated semantic sim-
ilarity metrics and hierarchical context into evaluation
frameworks may yield deeper insights into model per-
formance in ontology-driven tasks. This work pro-
vides a foundation for developing scalable and accu-
rate models for ontology annotation in specialized do-
mains like biomedical sciences.
ACKNOWLEDGEMENTS
This work is funded by a CAREER award (#1942727)
from the Division of Biological Infrastructure at the
National Science Foundation, USA.
REFERENCES
Bender, E. M., Gebru, T., McMillan-Major, A., and
Shmitchell, S. (2021). On the dangers of stochastic
parrots: Can language models be too big? In Pro-
ceedings of the 2021 ACM conference on fairness, ac-
countability, and transparency, pages 610–623.
Boguslav, M. R., Hailu, N. D., Bada, M., Baumgartner,
W. A., and Hunter, L. E. (2021). Concept recognition
as a machine translation problem. BMC bioinformat-
ics, 22(1):1–39.
Brown, T. B. (2020). Language models are few-shot learn-
ers.
Casteleiro, M. A., Demetriou, G., Read, W., Prieto, M. J. F.,
Maroto, N., Fernandez, D. M., Nenadic, G., Klein,
J., Keane, J., and Stevens, R. (2018). Deep learning
meets ontologies: experiments to anchor the cardio-
vascular disease ontology in the biomedical literature.
Journal of biomedical semantics, 9(1):13.
Dahdul, W., Dececchi, T. A., Ibrahim, N., Lapp, H., and
Mabee, P. (2015). Moving the mountain: analysis of
the effort required to transform comparative anatomy
into computable anatomy. Database, 2015.
Devkota, P., Mohanty, S., and Manda, P. (2022a). Knowl-
edge of the ancestors: Intelligent ontology-aware an-
notation of biological literature using semantic simi-
larity.
Devkota, P., Mohanty, S., and Manda, P. (2023). Ontology-
powered boosting for improved recognition of on-
tology concepts from biological literature [ontology-
powered boosting for improved recognition of ontol-
ogy concepts from biological literature]. In 16th In-
ternational Joint Conference on Biomedical Engineer-
ing Systems and Technologies (BIOSTEC 2023), vol-
ume 3.
Devkota, P., Mohanty, S. D., and Manda, P. (2022b). A
gated recurrent unit based architecture for recognizing
ontology concepts from biological literature. BioData
Mining, 15(1):1–23.
Lample, G., Ballesteros, M., Subramanian, S., Kawakami,
K., and Dyer, C. (2016). Neural architectures for
named entity recognition.
Manda, P., Beasley, L., and Mohanty, S. (2018). Taking
a dive: Experiments in deep learning for automatic
ontology-based annotation of scientific literature.
Manda, P., SayedAhmed, S., and Mohanty, S. D. (2020).
Automated ontology-based annotation of scientific lit-
erature using deep learning. In Proceedings of The
International Workshop on Semantic Big Data, SBD
’20, New York, NY, USA. Association for Computing
Machinery.
Pesquita, C., Faria, D., Falcao, A. O., Lord, P., and Couto,
F. M. (2009). Semantic similarity in biomedical on-
tologies. PLoS computational biology, 5(7).
Pratik, D., Somya, D. M., and Prashanti, M. (2023). Im-
proving the evaluation of nlp approaches for scien-
tific text annotation with ontology embedding-based
semantic similarity metrics. In Proceedings of the
20th International Conference on Natural Language
Processing (ICON), pages 516–522.
Strubell, E., Ganesh, A., and McCallum, A. (2020). Energy
and policy considerations for modern deep learning
research.
Vaswani, A. (2017). Attention is all you need.
BIOINFORMATICS 2025 - 16th International Conference on Bioinformatics Models, Methods and Algorithms
494