7 CONCLUSION
Leveraging KGs to enhance LLMs is a promising ap-
proach to increasing the accuracy of responses and re-
ducing hallucinations and incorrect facts. In this doc-
ument, a system is introduced to retrieve academic lit-
erature information through natural language queries
and responses. After the evaluation of the solution, it
can be concluded that the proposed approach halluci-
nates less frequently than an LLM without KGs.
For future work, different prompt templates could
be tried and compared, namely to improve the gen-
eration of SPARQL code. We also envision the ex-
pansion of the DBLP knowledge base to include the
abstract or even the body of the publication, which
would allow the LLM to answer queries that require
reasoning about the content of publications. An-
other possibility would be the integration of knowl-
edge bases of academic publications in fields other
than computer science. Additionally, we could ana-
lyze and assess the limitations associated with having
intermediate steps, as these introduce multiple points
of failure. By identifying and evaluating the error po-
tential at each stage, we can pinpoint the most critical
step and focus our efforts on improving the overall
system.
REFERENCES
Agrawal, G., Kumarage, T., Alghami, Z., and Liu, H.
(2023). Can knowledge graphs reduce hallucinations
in llms? : A survey. CoRR, abs/2311.07914.
Brate, R., Dang, M.-H., Hoppe, F., He, Y., Mero
˜
no-Pe
˜
nuela,
A., and Sadashivaiah, V. (2022). Improving Language
Model Predictions via Prompts Enriched with Knowl-
edge Graphs ?. In DL4KG@ ISWC2022, Hangzhou,
China.
Emsley, R. (2023). Chatgpt: these are not hallucinations –
they’re fabrications and falsifications. Schizophrenia,
9(1):52.
Goddard, J. (2023). Hallucinations in chatgpt: A cautionary
tale for biomedical researchers. The American Journal
of Medicine, 136(11):1059–1060.
Harper, F. M. and Konstan, J. A. (2016). The movielens
datasets: History and context. ACM Trans. Interact.
Intell. Syst., 5(4):19:1–19:19.
Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., Ishii, E.,
Bang, Y. J., Madotto, A., and Fung, P. (2023). Survey
of hallucination in natural language generation. ACM
Comput. Surv., 55(12).
Ley, M. (2002). The dblp computer science bibliography:
Evolution, research issues, perspectives. In Laender,
A. H. F. and Oliveira, A. L., editors, String Processing
and Information Retrieval, pages 1–10, Berlin, Hei-
delberg. Springer Berlin Heidelberg.
Martino, A., Iannelli, M., and Truong, C. (2023). Knowl-
edge injection to counter large language model (llm)
hallucination. In Pesquita, C., Skaf-Molli, H.,
Efthymiou, V., Kirrane, S., Ngonga, A., Collarana, D.,
Cerqueira, R., Alam, M., Trojahn, C., and Hertling,
S., editors, The Semantic Web: ESWC 2023 Satel-
lite Events, pages 182–185, Cham. Springer Nature
Switzerland.
Pan, J. Z., Razniewski, S., Kalo, J.-C., Singhania, S., Chen,
J., Dietze, S., Jabeen, H., Omeliyanenko, J., Zhang,
W., Lissandrini, M., Biswas, R., de Melo, G., Boni-
fati, A., Vakaj, E., Dragoni, M., and Graux, D. (2023).
Large language models and knowledge graphs: Op-
portunities and challenges.
Pan, J. Z., Vetere, G., Gomez-Perez, J. M., and Wu,
H. (2017). Exploiting Linked Data and Knowledge
Graphs in Large Organisations. Springer Publishing
Company, Incorporated, 1st edition.
Wang, C., Liu, X., Yue, Y., Tang, X., Zhang, T., Jiayang,
C., Yao, Y., Gao, W., Hu, X., Qi, Z., Wang, Y.,
Yang, L., Wang, J., Xie, X., Zhang, Z., and Zhang, Y.
(2023). Survey on factuality in large language models:
Knowledge, retrieval and domain-specificity. CoRR,
abs/2310.07521.
Enhancing LLMs with Knowledge Graphs for Academic Literature Retrieval
347