
This paper presents the development of a RAG-based
Chatbot for WhatsApp, designed to support users of
the PDSA-RS platform. Leveraging the combined
strengths of LLMs and retrieval systems, the chat-
bot aims to deliver accurate, contextually aware in-
teractions, enhancing user engagement and improving
understanding of the platform’s functionalities (Gao
et al., 2018).
The paper is structured as follows: Section 2 pro-
vides an overview of the key background concepts rel-
evant to the research and the proposed approach. Sec-
tion 3 introduces the methodology for developing the
Chatbot for the Rio Grande do Sul Animal Health De-
fense Platform (PDSA-RS), specifically designed to
provide support through WhatsApp. The simulation
of the proposed solution is detailed in Section 4. Fi-
nally, Section 5 presents the conclusions of this work.
2 BACKGROUND
2.1 Chatbots
Chatbots are software applications engineered to
mimic human conversation, predominantly through
text-based interfaces. Their development has seen
substantial advancements, transitioning from basic
rule-based systems to sophisticated AI-powered con-
versational agents (Pantano and Pizzi, 2023).
The inception of chatbots dates back to the mid-
20th century. A notable milestone was achieved in
1966 with the creation of ELIZA by Joseph Weizen-
baum. ELIZA utilized pattern matching and substitu-
tion techniques to simulate dialogue, setting a founda-
tional precedent for subsequent advancements in nat-
ural language processing (NLP) and conversational
AI (Weizenbaum, 1966).
The 2010s witnessed a pivotal transition towards
AI-driven chatbots. The emergence of machine learn-
ing and NLP technologies significantly enhanced
chatbots’ ability to comprehend and generate human-
like responses. Platforms such as IBM Watson, Mi-
crosoft Bot Framework, and Google’s Dialogflow
provided developers with robust tools to create more
intelligent and versatile chatbots (Wolf et al., 2019).
The 2020s have brought further innovations in
conversational AI, with chatbots becoming increas-
ingly context-aware and capable of managing intri-
cate queries. Large Language Models (LLMs) have
been instrumental in this progress, enabling chatbots
to understand and produce text in a manner akin to
human communication. Companies are progressively
integrating chatbots into diverse applications, ranging
from customer service to personal assistants and edu-
cational tools (Radford et al., 2019).
The evolution of chatbots, from simple rule-based
systems to advanced AI-driven softwares, has been
marked by continuous innovation. Driven by ad-
vancements in NLP and AI, chatbots have become in-
tegral to modern technology. As research progresses,
chatbots are set to further transform user experiences
across various domains (Adamopoulou and Moussi-
ades, 2020).
2.2 WhatsApp
WhatsApp is a popular messaging app that allows
users to send text messages, make voice and video
calls, and share media files. It was founded in 2009
by Brian Acton and Jan Koum and was acquired by
Facebook in 2014. WhatsApp has become one of the
most widely used communication tools globally, with
over 2 billion users.
WhatsApp offers a variety of features that make
it a versatile communication tool, including text mes-
saging, voice and video calls, media sharing, status
updates, and end-to-end encryption. Its widespread
adoption and continuous updates make it a reliable
choice for both personal and professional communi-
cation (Times, 2021).
2.3 Large Language Models
Large Language Models (LLMs) represent a signifi-
cant advancement in the field of artificial intelligence,
particularly in natural language processing. These
models have evolved rapidly, driven by innovations
in deep learning and the availability of vast amounts
of text data (Chang et al., 2024).
The 2010s saw the introduction of transformative
models like Word2Vec and GloVe, which focused on
word embeddings. The real breakthrough came with
the introduction of the Transformer architecture by
Vaswani et al. in their 2017 paper ”Attention is All
You Need.” This architecture, based on self-attention
mechanisms, allowed for more efficient processing of
sequential data, paving the way for larger and more
complex language models (Singh, 2024).
In 2018, Google introduced BERT (Bidirectional
Encoder Representations from Transformers), which
revolutionized the field by demonstrating state-of-
the-art performance on a wide range of natural lan-
guage processing tasks. The early 2020s witnessed a
proliferation of LLMs, with models like RoBERTa,
T5 (Text-to-Text Transfer Transformer), and others
building on BERT’s success. Notably, the introduc-
tion of models like ChatGPT by OpenAI showcased
ICEIS 2025 - 27th International Conference on Enterprise Information Systems
1078