loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Paper Unlock

Authors: Mariana Fidalgo Fernandes 1 and Plinio Moreno 2

Affiliations: 1 Instituto Superior Técnico, Universidade de Lisboa, 1049-001 Lisboa, Portugal ; 2 Institute for Systems and Robotics, Instituto Superior Técnico, Universidade de Lisboa, Torre Norte Piso 7, 1049-001 Lisboa, Portugal

Keyword(s): Natural Language Processing, Deep Learning, Machine Translation, Transformer, Attention Mechanism.

Abstract: Generative pre-trained transformers belong to the breakthroughs in Natural Language Processing (NLP), allowing Human-Robot Interactions (e.g. the creation of an open-domain chatbot). However, a substantial amount of research and available data are in English, causing low-resourced languages to be overlooked. This work addresses this problem for European Portuguese with two options: (i) Translation of the sentences before and after using the model fine-tuned on an English-based dataset, (ii) Translation of the English-based dataset to Portuguese and then fine-tune this model on it. We rely on the DialoGPT (dialogue generative pre-trained transformer), a tunable neural conversational answer generation model that learns the basic skills to conduct a dialogue. We use two sources of evaluation: (i) Metrics for text generation based on uncertainty (i.e. perplexity), and similarity between sentences (i.e. BLEU, METEOR and ROUGE) and (ii) Human-based evaluation of the sentences. The translat ion of sentences before and after of the modified DialoGPT model, using the Daily Dialogue dataset led to the best results. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 18.188.171.59

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Fernandes, M. F. and Moreno, P. (2022). Open-domain Conversational Agent based on Pre-trained Transformers for Human-Robot Interaction. In Proceedings of the 3rd International Conference on Deep Learning Theory and Applications - DeLTA; ISBN 978-989-758-584-5; ISSN 2184-9277, SciTePress, pages 168-175. DOI: 10.5220/0011300800003277

@conference{delta22,
author={Mariana Fidalgo Fernandes and Plinio Moreno},
title={Open-domain Conversational Agent based on Pre-trained Transformers for Human-Robot Interaction},
booktitle={Proceedings of the 3rd International Conference on Deep Learning Theory and Applications - DeLTA},
year={2022},
pages={168-175},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0011300800003277},
isbn={978-989-758-584-5},
issn={2184-9277},
}

TY - CONF

JO - Proceedings of the 3rd International Conference on Deep Learning Theory and Applications - DeLTA
TI - Open-domain Conversational Agent based on Pre-trained Transformers for Human-Robot Interaction
SN - 978-989-758-584-5
IS - 2184-9277
AU - Fernandes, M.
AU - Moreno, P.
PY - 2022
SP - 168
EP - 175
DO - 10.5220/0011300800003277
PB - SciTePress