
Lloyd, S. (1982). Least squares quantization in pcm. IEEE
Transactions on Information Theory, 28(2):129–137.
Loshchilov, I. and Hutter, F. (2017). Decoupled weight de-
cay regularization. arXiv preprint arXiv:1711.05101.
Mihalcea, R. and Tarau, P. (2004). TextRank: Bringing or-
der into text. In Lin, D. and Wu, D., editors, Pro-
ceedings of the 2004 Conference on Empirical Meth-
ods in Natural Language Processing, pages 404–411,
Barcelona, Spain. Association for Computational Lin-
guistics.
Miller, D. (2019). Leveraging bert for extractive text sum-
marization on lectures.
Nallapati, R., Zhai, F., and Zhou, B. (2017). Summarunner:
A recurrent neural network based sequence model for
extractive summarization of documents. In Proceed-
ings of the Thirty-First AAAI Conference on Artificial
Intelligence (AAAI), pages 3075–3081.
Narayan, S., Cohen, S. B., and Lapata, M. (2018). Ex-
treme summarization (xsum). In Proceedings of the
2018 Conference on Empirical Methods in Natural
Language Processing, pages 931–936.
OpenAI (2024). Gpt-4 technical report.
Page, L., Brin, S., Motwani, R., and Winograd, T. (1999).
The pagerank citation ranking : Bringing order to the
web. In The Web Conference.
Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S.,
Matena, M., Zhou, Y., Li, W., and Liu, P. J. (2020a).
Exploring the limits of transfer learning with a unified
text-to-text transformer. Journal of Machine Learning
Research, 21(140):1–67.
Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S.,
Matena, M., Zhou, Y., Li, W., and Liu, P. J. (2020b).
Exploring the limits of transfer learning with a unified
text-to-text transformer. Journal of Machine Learning
Research, 21(140):1–67.
Scialom, T. et al. (2020). Mlsum: Multilingual summariza-
tion dataset. In Proceedings of the 2020 Conference
on Empirical Methods in Natural Language Process-
ing, pages 2146–2161.
Straka, M., Mediankin, N., Kocmi, T.,
ˇ
Zabokrtsk
´
y, Z.,
Hude
ˇ
cek, V., and Haji
ˇ
c, J. (2018). SumeCzech: Large
Czech news-based summarization dataset. In Pro-
ceedings of the Eleventh International Conference on
Language Resources and Evaluation (LREC 2018),
Miyazaki, Japan. European Language Resources As-
sociation (ELRA).
Straka, M. and Strakov
´
a, J. (2018). Rougeraw: Language-
agnostic evaluation for summarization. Proceedings
of the International Conference on Computational
Linguistics.
Vaswani, A., Bengio, S., Brevdo, E., Chollet, F., Gomez,
A. N., Gouws, S., Jones, L., Kaiser, L., Kalchbrenner,
N., Parmar, N., Sepassi, R., Shazeer, N., and Uszkor-
eit, J. (2018). Tensor2tensor for neural machine trans-
lation. CoRR, abs/1803.07416.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones,
L., Gomez, A. N., Kaiser, L. u., and Polosukhin,
I. (2017). Attention is all you need. In Guyon,
I., Luxburg, U. V., Bengio, S., Wallach, H., Fer-
gus, R., Vishwanathan, S., and Garnett, R., editors,
Advances in Neural Information Processing Systems,
volume 30. Curran Associates, Inc.
Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue,
C., Moi, A., Cistac, P., Rault, T., Louf, R., Funtow-
icz, M., Davison, J., Shleifer, S., von Platen, P., Ma,
C., Jernite, Y., Plu, J., Xu, C., Scao, T. L., Gugger,
S., Drame, M., Lhoest, Q., and Rush, A. M. (2020).
Transformers: State-of-the-art natural language pro-
cessing. In Proceedings of the 2020 Conference on
Empirical Methods in Natural Language Processing:
System Demonstrations, pages 38–45, Online. Asso-
ciation for Computational Linguistics.
Xue, L., Constant, N., Roberts, A., Kale, M., Al-Rfou,
R., Siddhant, A., Barua, A., and Raffel, C. (2021a).
mC4: A massively multilingual cleaned crawl corpus.
In Proceedings of the 2021 Conference on Empirical
Methods in Natural Language Processing (EMNLP),
pages 7517–7532, Online and Punta Cana, Dominican
Republic. Association for Computational Linguistics.
Xue, L., Constant, N., Roberts, A., Kale, M., Al-Rfou,
R., Siddhant, A., Barua, A., and Raffel, C. (2021b).
mT5: A massively multilingual pre-trained text-to-
text transformer. In Toutanova, K., Rumshisky,
A., Zettlemoyer, L., Hakkani-Tur, D., Beltagy, I.,
Bethard, S., Cotterell, R., Chakraborty, T., and Zhou,
Y., editors, Proceedings of the 2021 Conference of the
North American Chapter of the Association for Com-
putational Linguistics: Human Language Technolo-
gies, pages 483–498, Online. Association for Compu-
tational Linguistics.
Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., and Hovy,
E. (2016). Hierarchical attention networks for docu-
ment classification. In Proceedings of the 2016 Con-
ference of the North American Chapter of the Associa-
tion for Computational Linguistics: Human Language
Technologies, pages 1480–1489.
Zhang, J., Zhao, Y., Saleh, M., and Liu, P. J. (2019). Pe-
gasus: Pre-training with extracted gap-sentences for
abstractive summarization.
ICAART 2025 - 17th International Conference on Agents and Artificial Intelligence
804